var/home/core/zuul-output/0000755000175000017500000000000015110722410014516 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110743671015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005627374515110743662017725 0ustar rootrootNov 24 00:24:56 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 00:24:56 crc restorecon[4747]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:56 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 00:24:57 crc restorecon[4747]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 00:24:57 crc kubenswrapper[4888]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.963693 4888 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972896 4888 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972948 4888 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972958 4888 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972968 4888 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972976 4888 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972986 4888 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.972996 4888 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973007 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973016 4888 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973024 4888 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973032 4888 feature_gate.go:330] unrecognized feature gate: Example Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973040 4888 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973048 4888 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973056 4888 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973064 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973077 4888 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973091 4888 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973100 4888 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973109 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973119 4888 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973130 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973141 4888 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973150 4888 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973159 4888 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973168 4888 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973176 4888 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973185 4888 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973193 4888 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973200 4888 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973208 4888 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973216 4888 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973224 4888 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973234 4888 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973242 4888 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973250 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973259 4888 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973267 4888 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973276 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973285 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973293 4888 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973302 4888 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973310 4888 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973318 4888 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973327 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973335 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973344 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973353 4888 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973362 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973373 4888 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973382 4888 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973390 4888 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973398 4888 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973408 4888 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973419 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973427 4888 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973437 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973445 4888 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973454 4888 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973462 4888 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973472 4888 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973481 4888 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973490 4888 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973500 4888 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973509 4888 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973518 4888 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973526 4888 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973534 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973548 4888 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973558 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973567 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.973576 4888 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974571 4888 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974596 4888 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974617 4888 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974630 4888 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974642 4888 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974652 4888 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974665 4888 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974678 4888 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974687 4888 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974697 4888 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974707 4888 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974720 4888 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974729 4888 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974739 4888 flags.go:64] FLAG: --cgroup-root="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974748 4888 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974757 4888 flags.go:64] FLAG: --client-ca-file="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974766 4888 flags.go:64] FLAG: --cloud-config="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974775 4888 flags.go:64] FLAG: --cloud-provider="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974784 4888 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974795 4888 flags.go:64] FLAG: --cluster-domain="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974804 4888 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974844 4888 flags.go:64] FLAG: --config-dir="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974854 4888 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974865 4888 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974878 4888 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974887 4888 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974897 4888 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974906 4888 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974916 4888 flags.go:64] FLAG: --contention-profiling="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974926 4888 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974935 4888 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974945 4888 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974954 4888 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974966 4888 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974976 4888 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974985 4888 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.974995 4888 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975004 4888 flags.go:64] FLAG: --enable-server="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975012 4888 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975027 4888 flags.go:64] FLAG: --event-burst="100" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975036 4888 flags.go:64] FLAG: --event-qps="50" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975046 4888 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975055 4888 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975064 4888 flags.go:64] FLAG: --eviction-hard="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975076 4888 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975085 4888 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975094 4888 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975104 4888 flags.go:64] FLAG: --eviction-soft="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975114 4888 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975123 4888 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975132 4888 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975141 4888 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975150 4888 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975159 4888 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975168 4888 flags.go:64] FLAG: --feature-gates="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975179 4888 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975189 4888 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975198 4888 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975209 4888 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975218 4888 flags.go:64] FLAG: --healthz-port="10248" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975232 4888 flags.go:64] FLAG: --help="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975242 4888 flags.go:64] FLAG: --hostname-override="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975252 4888 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975262 4888 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975272 4888 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975281 4888 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975291 4888 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975300 4888 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975308 4888 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975319 4888 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975328 4888 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975337 4888 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975347 4888 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975357 4888 flags.go:64] FLAG: --kube-reserved="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975366 4888 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975376 4888 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975385 4888 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975394 4888 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975403 4888 flags.go:64] FLAG: --lock-file="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975412 4888 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975422 4888 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975432 4888 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975457 4888 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975467 4888 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975476 4888 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975484 4888 flags.go:64] FLAG: --logging-format="text" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975494 4888 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975504 4888 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975513 4888 flags.go:64] FLAG: --manifest-url="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975522 4888 flags.go:64] FLAG: --manifest-url-header="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975535 4888 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975545 4888 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975557 4888 flags.go:64] FLAG: --max-pods="110" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975566 4888 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975575 4888 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975586 4888 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975596 4888 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975605 4888 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975615 4888 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975624 4888 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975649 4888 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975659 4888 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975668 4888 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975678 4888 flags.go:64] FLAG: --pod-cidr="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975687 4888 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975700 4888 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975710 4888 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975720 4888 flags.go:64] FLAG: --pods-per-core="0" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975729 4888 flags.go:64] FLAG: --port="10250" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975738 4888 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975747 4888 flags.go:64] FLAG: --provider-id="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975757 4888 flags.go:64] FLAG: --qos-reserved="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975766 4888 flags.go:64] FLAG: --read-only-port="10255" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975775 4888 flags.go:64] FLAG: --register-node="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975784 4888 flags.go:64] FLAG: --register-schedulable="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975793 4888 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975832 4888 flags.go:64] FLAG: --registry-burst="10" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975842 4888 flags.go:64] FLAG: --registry-qps="5" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975852 4888 flags.go:64] FLAG: --reserved-cpus="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975862 4888 flags.go:64] FLAG: --reserved-memory="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975875 4888 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975884 4888 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975894 4888 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975903 4888 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975913 4888 flags.go:64] FLAG: --runonce="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975923 4888 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975934 4888 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975975 4888 flags.go:64] FLAG: --seccomp-default="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975988 4888 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.975998 4888 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976008 4888 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976019 4888 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976029 4888 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976039 4888 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976048 4888 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976057 4888 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976067 4888 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976078 4888 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976088 4888 flags.go:64] FLAG: --system-cgroups="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976098 4888 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976115 4888 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976124 4888 flags.go:64] FLAG: --tls-cert-file="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976134 4888 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976146 4888 flags.go:64] FLAG: --tls-min-version="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976155 4888 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976165 4888 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976174 4888 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976183 4888 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976193 4888 flags.go:64] FLAG: --v="2" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976205 4888 flags.go:64] FLAG: --version="false" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976218 4888 flags.go:64] FLAG: --vmodule="" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976229 4888 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.976239 4888 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976500 4888 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976513 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976523 4888 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976533 4888 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976544 4888 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976555 4888 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976567 4888 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976579 4888 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976589 4888 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976599 4888 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976608 4888 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976618 4888 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976627 4888 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976635 4888 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976643 4888 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976651 4888 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976659 4888 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976668 4888 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976677 4888 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976686 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976695 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976704 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976713 4888 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976721 4888 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976730 4888 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976738 4888 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976747 4888 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976756 4888 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976767 4888 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976780 4888 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976789 4888 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976798 4888 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976806 4888 feature_gate.go:330] unrecognized feature gate: Example Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976838 4888 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976846 4888 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976854 4888 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976863 4888 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976872 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976882 4888 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976891 4888 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976900 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976907 4888 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976916 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976923 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976932 4888 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976939 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976948 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976956 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976964 4888 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976972 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976980 4888 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976988 4888 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.976996 4888 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977004 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977012 4888 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977020 4888 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977028 4888 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977036 4888 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977044 4888 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977055 4888 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977065 4888 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977074 4888 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977084 4888 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977093 4888 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977102 4888 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977111 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977119 4888 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977126 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977135 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977143 4888 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.977152 4888 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.978846 4888 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.995638 4888 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.995735 4888 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995916 4888 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995960 4888 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995969 4888 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995977 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995986 4888 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.995994 4888 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996000 4888 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996006 4888 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996013 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996018 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996025 4888 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996034 4888 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996043 4888 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996051 4888 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996059 4888 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996258 4888 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996264 4888 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996271 4888 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996279 4888 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996287 4888 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996294 4888 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996302 4888 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996310 4888 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996317 4888 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996323 4888 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996330 4888 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996336 4888 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996342 4888 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996352 4888 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996361 4888 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996368 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996376 4888 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996382 4888 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996389 4888 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996397 4888 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996404 4888 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996411 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996416 4888 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996422 4888 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996428 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996433 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996439 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996444 4888 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996449 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996454 4888 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996460 4888 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996465 4888 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996471 4888 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996476 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996481 4888 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996487 4888 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996493 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996499 4888 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996504 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996509 4888 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996515 4888 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996520 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996525 4888 feature_gate.go:330] unrecognized feature gate: Example Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996530 4888 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996536 4888 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996541 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996546 4888 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996552 4888 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996557 4888 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996562 4888 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996569 4888 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996575 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996582 4888 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996587 4888 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996593 4888 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996599 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 00:24:57 crc kubenswrapper[4888]: I1124 00:24:57.996609 4888 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996791 4888 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996800 4888 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996839 4888 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996850 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996857 4888 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996863 4888 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996870 4888 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996877 4888 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996884 4888 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996891 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996898 4888 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996906 4888 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996916 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996924 4888 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996931 4888 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996938 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996944 4888 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996951 4888 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996957 4888 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996964 4888 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996970 4888 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996978 4888 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996984 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996990 4888 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.996996 4888 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997003 4888 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997010 4888 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997017 4888 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997024 4888 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997030 4888 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997037 4888 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997043 4888 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997050 4888 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997056 4888 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997064 4888 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997071 4888 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997078 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997083 4888 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997272 4888 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997278 4888 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997283 4888 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997290 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997296 4888 feature_gate.go:330] unrecognized feature gate: Example Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997302 4888 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997311 4888 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997318 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997325 4888 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997335 4888 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997342 4888 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997349 4888 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997361 4888 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997376 4888 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997385 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997393 4888 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997400 4888 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997409 4888 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997419 4888 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997427 4888 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 00:24:57 crc kubenswrapper[4888]: W1124 00:24:57.997434 4888 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997441 4888 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997449 4888 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997459 4888 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997467 4888 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997474 4888 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997480 4888 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997486 4888 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997498 4888 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997504 4888 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997510 4888 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997515 4888 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:57.997524 4888 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:57.997535 4888 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:57.997866 4888 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.010073 4888 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.010198 4888 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.012658 4888 server.go:997] "Starting client certificate rotation" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.012690 4888 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.014173 4888 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 06:41:32.244567248 +0000 UTC Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.014278 4888 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 510h16m34.230292599s for next certificate rotation Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.049931 4888 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.052773 4888 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.071398 4888 log.go:25] "Validated CRI v1 runtime API" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.105910 4888 log.go:25] "Validated CRI v1 image API" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.107991 4888 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.116501 4888 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-00-20-30-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.116536 4888 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.141386 4888 manager.go:217] Machine: {Timestamp:2025-11-24 00:24:58.138241594 +0000 UTC m=+0.720925768 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:99fa3412-cc6b-4ec6-b00c-fbecc545cd5b BootID:2aca63cc-91d1-4463-bed3-9c285b4b381e Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:93:3d:06 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:93:3d:06 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:78:1e:c4 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b3:fc:5d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:75:2a:51 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:65:37:97 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:87:47:39 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:22:5e:6d:97:0b:db Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:12:bb:72:00:44:f0 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.141765 4888 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.142019 4888 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.143368 4888 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.143646 4888 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.143699 4888 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.144075 4888 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.144096 4888 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.144676 4888 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.144718 4888 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.144987 4888 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.145093 4888 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.151142 4888 kubelet.go:418] "Attempting to sync node with API server" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.151173 4888 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.151213 4888 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.151232 4888 kubelet.go:324] "Adding apiserver pod source" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.151249 4888 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.157214 4888 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.158470 4888 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:58.159214 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.159361 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:58.159205 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.159526 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.162585 4888 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164426 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164453 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164462 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164470 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164483 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164491 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164498 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164509 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164525 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164533 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164554 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.164562 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.165767 4888 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.166364 4888 server.go:1280] "Started kubelet" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.168613 4888 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 00:24:58 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.175958 4888 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.176072 4888 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.168445 4888 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.178042 4888 server.go:460] "Adding debug handlers to kubelet server" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.178600 4888 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.178650 4888 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.178710 4888 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 17:58:06.097295557 +0000 UTC Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.178862 4888 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1193h33m7.918443702s for next certificate rotation Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.179025 4888 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.179040 4888 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.179054 4888 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.179133 4888 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.180366 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="200ms" Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:58.180479 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.180855 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.181197 4888 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.23:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ac9a12debf484 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 00:24:58.166326404 +0000 UTC m=+0.749010448,LastTimestamp:2025-11-24 00:24:58.166326404 +0000 UTC m=+0.749010448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.185162 4888 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.185200 4888 factory.go:55] Registering systemd factory Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.185213 4888 factory.go:221] Registration of the systemd container factory successfully Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.186283 4888 factory.go:153] Registering CRI-O factory Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.186305 4888 factory.go:221] Registration of the crio container factory successfully Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.186338 4888 factory.go:103] Registering Raw factory Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.186354 4888 manager.go:1196] Started watching for new ooms in manager Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.192495 4888 manager.go:319] Starting recovery of all containers Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196481 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196550 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196562 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196571 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196581 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196590 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196601 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196611 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196624 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196634 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196644 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196654 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196663 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196673 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196685 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196695 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196706 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196715 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196724 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196733 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196742 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196751 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196760 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196771 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196780 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196792 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196822 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196836 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196846 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196858 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196868 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196879 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196942 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196953 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196962 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196973 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196982 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.196996 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197006 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197018 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197029 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197039 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197051 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197061 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197071 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197083 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197094 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197105 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197115 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197127 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197136 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197147 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197162 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197173 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197184 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197195 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197206 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197236 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197246 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197258 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197268 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197279 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197290 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197303 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197314 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197324 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197333 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197344 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197356 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197366 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197377 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197387 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197397 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197407 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197417 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197426 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.197437 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200215 4888 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200269 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200294 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200304 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200314 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200325 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200413 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200424 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200434 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200445 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200455 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200466 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200499 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200510 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200520 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200529 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200540 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200554 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200563 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200574 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200583 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200593 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200602 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200611 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200619 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200630 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200642 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200652 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200667 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200678 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200688 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200699 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200717 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200770 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200785 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200805 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200832 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200845 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200856 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200865 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200875 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200885 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200895 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200905 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200915 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200971 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200982 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.200990 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201008 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201017 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201025 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201034 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201043 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201054 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201064 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201073 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201083 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201094 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201104 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201114 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201124 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201134 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201143 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201153 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201164 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201181 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201196 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201208 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201221 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201233 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201244 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201254 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201264 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201273 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201282 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201291 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201301 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201311 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201320 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201330 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201342 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201354 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201378 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201392 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201406 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201420 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201432 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201445 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201458 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201548 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201563 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201576 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201592 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201605 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201616 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201628 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201641 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201655 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201669 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201681 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201695 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201706 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201720 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201734 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201746 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201757 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201766 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201775 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201786 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201800 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201833 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201844 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201856 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201868 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201879 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201893 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201904 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201914 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201924 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201937 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201947 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.201956 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202005 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202016 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202027 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202039 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202048 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202066 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202077 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202088 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.202098 4888 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.203480 4888 reconstruct.go:97] "Volume reconstruction finished" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.203545 4888 reconciler.go:26] "Reconciler: start to sync state" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.212627 4888 manager.go:324] Recovery completed Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.225354 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.234004 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.234293 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.234315 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.237642 4888 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.237723 4888 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.237786 4888 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.239523 4888 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.242766 4888 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.242906 4888 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.243171 4888 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:58.243828 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.244040 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.244162 4888 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.253630 4888 policy_none.go:49] "None policy: Start" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.254761 4888 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.254835 4888 state_mem.go:35] "Initializing new in-memory state store" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.279253 4888 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.324328 4888 manager.go:334] "Starting Device Plugin manager" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.324417 4888 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.324430 4888 server.go:79] "Starting device plugin registration server" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.325498 4888 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.325535 4888 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.325935 4888 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.326117 4888 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.326131 4888 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.334824 4888 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.345400 4888 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.345501 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.346847 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.346882 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.346890 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347000 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347325 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347390 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347708 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.347777 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348027 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348179 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348223 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348430 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.348465 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349116 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349128 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349156 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349169 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349291 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349414 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349445 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.349973 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350026 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350041 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350253 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350256 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350353 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350387 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.350452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351203 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351228 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351237 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351340 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351355 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351363 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351508 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.351528 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.352179 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.352200 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.352209 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.381531 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="400ms" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406684 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406731 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406751 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406768 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406789 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406838 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406864 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.406943 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407436 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407498 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407550 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407633 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407714 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.407762 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.427563 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.430234 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.430286 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.430305 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.430343 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.430962 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.509644 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.510055 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.510974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.510624 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.511507 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.511879 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.512255 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.512158 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.511769 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.512523 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.512862 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.513149 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.513548 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.513697 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.513733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.513911 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514066 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514162 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514209 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514289 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514338 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514413 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514555 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514554 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514649 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514688 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514720 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514780 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.514934 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.631075 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.633020 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.633082 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.633095 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.633130 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.633745 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.678662 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.686014 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.699417 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.716562 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: I1124 00:24:58.726194 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:24:58 crc kubenswrapper[4888]: W1124 00:24:58.737962 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ca224a99671405004394876fbc1ffc5c209c23d266a3730673919a5bf61f1450 WatchSource:0}: Error finding container ca224a99671405004394876fbc1ffc5c209c23d266a3730673919a5bf61f1450: Status 404 returned error can't find the container with id ca224a99671405004394876fbc1ffc5c209c23d266a3730673919a5bf61f1450 Nov 24 00:24:58 crc kubenswrapper[4888]: E1124 00:24:58.782763 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="800ms" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.034104 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.036422 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.036486 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.036506 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.036548 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.037442 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Nov 24 00:24:59 crc kubenswrapper[4888]: W1124 00:24:59.076553 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.076672 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:59 crc kubenswrapper[4888]: W1124 00:24:59.088178 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.088286 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:59 crc kubenswrapper[4888]: W1124 00:24:59.116606 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.116700 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.177436 4888 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.248871 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ca224a99671405004394876fbc1ffc5c209c23d266a3730673919a5bf61f1450"} Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.250595 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6d4ad47715f6ca3657c7c0c545307406bd99d726f6f3a5907b0d79c72914af23"} Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.251791 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f889b4235e8d6449b41387f1c2ee1dd25690ef8cf9fe1d7746510957f918667d"} Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.253265 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5da3f73049f3fdf7939d9e8955337af7b55c0db3759a4c04d0117829750a8650"} Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.254857 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a09ee9a275aa9138bad4142e32784cb99fc54c8d8d17e8a50a49783a92c21cf0"} Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.395030 4888 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.23:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ac9a12debf484 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 00:24:58.166326404 +0000 UTC m=+0.749010448,LastTimestamp:2025-11-24 00:24:58.166326404 +0000 UTC m=+0.749010448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 00:24:59 crc kubenswrapper[4888]: W1124 00:24:59.437229 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.437315 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.583849 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="1.6s" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.838507 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.839954 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.840002 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.840013 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:24:59 crc kubenswrapper[4888]: I1124 00:24:59.840040 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:24:59 crc kubenswrapper[4888]: E1124 00:24:59.840362 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.178021 4888 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.269037 4888 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8" exitCode=0 Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.269229 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.269235 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.271206 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.271255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.271276 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.272936 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.272999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.273021 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.273045 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.273188 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.275778 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.275837 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.275807 4888 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc" exitCode=0 Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.275862 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.276054 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.276740 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.277097 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.277127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.277145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.279323 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.279997 4888 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7" exitCode=0 Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.280126 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.280182 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.280848 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.280873 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.280882 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.281936 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.281959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.281968 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.283847 4888 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00" exitCode=0 Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.283914 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00"} Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.284027 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.285149 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.285202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:00 crc kubenswrapper[4888]: I1124 00:25:00.285225 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: W1124 00:25:01.074379 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:25:01 crc kubenswrapper[4888]: E1124 00:25:01.074498 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.178019 4888 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:25:01 crc kubenswrapper[4888]: E1124 00:25:01.184612 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="3.2s" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.293230 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.293236 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.293307 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.293320 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.296329 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.296378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.296390 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.303662 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.303749 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.303766 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.303782 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.306087 4888 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d" exitCode=0 Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.306191 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.306237 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.307193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.307225 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.307237 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.309171 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.309569 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.309722 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d"} Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.310351 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.310395 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.310411 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.311003 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.311263 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.311277 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.449896 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.451227 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.451284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.451296 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:01 crc kubenswrapper[4888]: I1124 00:25:01.451328 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:25:01 crc kubenswrapper[4888]: E1124 00:25:01.451907 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Nov 24 00:25:01 crc kubenswrapper[4888]: W1124 00:25:01.564228 4888 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Nov 24 00:25:01 crc kubenswrapper[4888]: E1124 00:25:01.564324 4888 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.314655 4888 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea" exitCode=0 Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.314841 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.314803 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea"} Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.316001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.316114 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.316136 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.318719 4888 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.318756 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.318752 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860"} Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.318915 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.318982 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319742 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319808 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319860 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319958 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319982 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.319991 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.320294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.320345 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.320364 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.640859 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.641082 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.642440 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.642480 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:02 crc kubenswrapper[4888]: I1124 00:25:02.642492 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326291 4888 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326338 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326579 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb"} Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326650 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7"} Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326667 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4"} Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.326679 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981"} Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.328015 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.328052 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:03 crc kubenswrapper[4888]: I1124 00:25:03.328063 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.336152 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d"} Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.336357 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.337842 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.337977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.338065 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.586024 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.586295 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.588317 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.588400 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.588419 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.653044 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.654950 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.655077 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.655197 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:04 crc kubenswrapper[4888]: I1124 00:25:04.655259 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.338441 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.339557 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.339608 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.339625 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.776806 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.777261 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.779167 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.779243 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:05 crc kubenswrapper[4888]: I1124 00:25:05.779261 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.264921 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.341753 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.343750 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.343854 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.343904 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.721701 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.722089 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.724102 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.724186 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.724206 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:06 crc kubenswrapper[4888]: I1124 00:25:06.886247 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:07 crc kubenswrapper[4888]: I1124 00:25:07.183965 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:07 crc kubenswrapper[4888]: I1124 00:25:07.345739 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:07 crc kubenswrapper[4888]: I1124 00:25:07.347593 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:07 crc kubenswrapper[4888]: I1124 00:25:07.347685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:07 crc kubenswrapper[4888]: I1124 00:25:07.347711 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:08 crc kubenswrapper[4888]: E1124 00:25:08.335102 4888 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.348766 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.350640 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.350723 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.350746 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.523796 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.524165 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.526401 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.526446 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:08 crc kubenswrapper[4888]: I1124 00:25:08.526463 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.218637 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.228038 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.351761 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.353434 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.353470 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.353484 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:09 crc kubenswrapper[4888]: I1124 00:25:09.359420 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:10 crc kubenswrapper[4888]: I1124 00:25:10.355075 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:10 crc kubenswrapper[4888]: I1124 00:25:10.358974 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:10 crc kubenswrapper[4888]: I1124 00:25:10.359122 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:10 crc kubenswrapper[4888]: I1124 00:25:10.359145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.357643 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.358591 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.358647 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.358658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.524674 4888 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.524779 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.925418 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.925674 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.927159 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.927239 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:11 crc kubenswrapper[4888]: I1124 00:25:11.927275 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:12 crc kubenswrapper[4888]: I1124 00:25:12.018757 4888 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 00:25:12 crc kubenswrapper[4888]: I1124 00:25:12.018932 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 00:25:12 crc kubenswrapper[4888]: I1124 00:25:12.026227 4888 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 00:25:12 crc kubenswrapper[4888]: I1124 00:25:12.026313 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.722805 4888 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.722917 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.890992 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.891346 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.891790 4888 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.891891 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.893641 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.893700 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.893719 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:16 crc kubenswrapper[4888]: I1124 00:25:16.897837 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.004704 4888 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.009354 4888 trace.go:236] Trace[312149270]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 00:25:02.160) (total time: 14848ms): Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[312149270]: ---"Objects listed" error: 14848ms (00:25:17.009) Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[312149270]: [14.848399156s] [14.848399156s] END Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.009396 4888 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.011391 4888 trace.go:236] Trace[855112095]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 00:25:04.511) (total time: 12499ms): Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[855112095]: ---"Objects listed" error: 12499ms (00:25:17.011) Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[855112095]: [12.49947502s] [12.49947502s] END Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.011486 4888 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.011828 4888 trace.go:236] Trace[1577892233]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 00:25:02.099) (total time: 14912ms): Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[1577892233]: ---"Objects listed" error: 14912ms (00:25:17.011) Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[1577892233]: [14.912399734s] [14.912399734s] END Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.011853 4888 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.012904 4888 trace.go:236] Trace[2060774447]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 00:25:06.539) (total time: 10472ms): Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[2060774447]: ---"Objects listed" error: 10472ms (00:25:17.012) Nov 24 00:25:17 crc kubenswrapper[4888]: Trace[2060774447]: [10.472958469s] [10.472958469s] END Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.012923 4888 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.014442 4888 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.015075 4888 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.164963 4888 apiserver.go:52] "Watching apiserver" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.168229 4888 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.168775 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.169303 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.170606 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.170843 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.170946 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.171124 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.171389 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.171438 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.171475 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.171775 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.173168 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.173910 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.173918 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.173994 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.174348 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.174839 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.177080 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.177334 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.177582 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.185167 4888 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.209864 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216204 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216257 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216282 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216307 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216345 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216366 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216387 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216408 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216432 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216456 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216478 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216503 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216531 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216555 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216579 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216609 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216637 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216651 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216665 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216721 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216751 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216779 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216828 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216899 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216928 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216933 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216960 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.216990 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217017 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217049 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217078 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217107 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217135 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217160 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217185 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217214 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217241 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217271 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217303 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217329 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217358 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217388 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217415 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217443 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217471 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217501 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217528 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217554 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217579 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217601 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217625 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217649 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217671 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217697 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217720 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217740 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217760 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217791 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217832 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217858 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217895 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217923 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217950 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217974 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217999 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218021 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218048 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218070 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218090 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218112 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218133 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218154 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218182 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218227 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218254 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218276 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218301 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218329 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218350 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218370 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218390 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218409 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218431 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218456 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218480 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218505 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218529 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218555 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218579 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218604 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218667 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218704 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218732 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218756 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218781 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218821 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218846 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218873 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218896 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218920 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218945 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218973 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219002 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219021 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219039 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219057 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219078 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219096 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219120 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219165 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219185 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219203 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219245 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219271 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219327 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219359 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219386 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219434 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219457 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217184 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219506 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219572 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219602 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219661 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219687 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219749 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219780 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219830 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219859 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219938 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219968 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220159 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220196 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220248 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220282 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220461 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220489 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220536 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220715 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220751 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.221386 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.221667 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222469 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222639 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222674 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222739 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222771 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222833 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222859 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222969 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222997 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223132 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223206 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223270 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223303 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223439 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223471 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.224042 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.230897 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232414 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232940 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232974 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233001 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233028 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233047 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233067 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233086 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233106 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233124 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233144 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233162 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233188 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233213 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233233 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233249 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233268 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233402 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233426 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233447 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233467 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233488 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233512 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233538 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233565 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233588 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233606 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233626 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233649 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233667 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233685 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233706 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233726 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233745 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233765 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233785 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233838 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233978 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234003 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234060 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234092 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234121 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234148 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234171 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234235 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234259 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234475 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234500 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234528 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234567 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234594 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234622 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234738 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234756 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234769 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.217428 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218231 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.218481 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219258 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219316 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.236268 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219343 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219443 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219538 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219581 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219591 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219645 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219685 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219726 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219824 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.219855 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220158 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220209 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.220403 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.221629 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.221745 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.221982 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222080 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222232 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222713 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222893 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.222920 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223062 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223113 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223174 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223512 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223821 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223938 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223983 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.223972 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.224194 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.224526 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.224776 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.225019 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.225774 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.225963 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.226373 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.226769 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.227067 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.228673 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.229209 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.229440 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.230117 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.230449 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.230567 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.230726 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231049 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231148 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231220 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231590 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231541 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231858 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.231874 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232246 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232314 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232346 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232453 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232529 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232570 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232675 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232943 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.232990 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233069 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233287 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233298 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233412 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233423 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233743 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.233953 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234118 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234414 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234553 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234642 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.234352 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.234946 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.235038 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.235236 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.235333 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.235480 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.235908 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.236842 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.236045 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.237165 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.236480 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.237207 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.237241 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.237332 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238053 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238101 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238522 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238533 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238603 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238633 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238880 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.238937 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.239248 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.240235 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.240452 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.242853 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.243045 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:17.742996735 +0000 UTC m=+20.325680959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.243749 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.244021 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.244570 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.244871 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.245631 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.245901 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.245972 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.245972 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.246127 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.246139 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.246184 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.246875 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:17.746849421 +0000 UTC m=+20.329533465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.247320 4888 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.247440 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.248202 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.248396 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.248431 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.248944 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.249051 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:17.74902454 +0000 UTC m=+20.331708764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.249742 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.250025 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.250068 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.250034 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.250486 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.250620 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.251113 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.251152 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.251307 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.251637 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.251377 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.252074 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.252115 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.253137 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.253359 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.253794 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.253955 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.257500 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.257723 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.257819 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.261934 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.262295 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.262508 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.262661 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.262752 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.265633 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.265969 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.266147 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.266183 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.266489 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.268035 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.268262 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.268299 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.268318 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.268408 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:17.76837961 +0000 UTC m=+20.351063664 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.272967 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.273213 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.273268 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.273371 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.274112 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.274248 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.274472 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.275280 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.275470 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.275951 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.276028 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.276028 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.276205 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.276226 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.276240 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.276311 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:17.776288586 +0000 UTC m=+20.358972630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.277231 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.278555 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.279600 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.281342 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.283799 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.284124 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.284144 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.284294 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.284691 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.285744 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.287352 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.287776 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.288171 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.288404 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.290524 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.294691 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.294927 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.294984 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.295439 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.295505 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.295533 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.296008 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.296455 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.296767 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.296899 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.299256 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.299360 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.300513 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.302367 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.302371 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.303113 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307154 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307227 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307245 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307306 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307777 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.307831 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.309799 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.310163 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.311040 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.318246 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.321212 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.322797 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.335750 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.335900 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336254 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336306 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336410 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336445 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336447 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336460 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336481 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336493 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336506 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336516 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336526 4888 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336536 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336545 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336556 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336568 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336578 4888 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336589 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336599 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336609 4888 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336644 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336667 4888 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336680 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336691 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336701 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336713 4888 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336722 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336733 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336744 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336757 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336767 4888 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336777 4888 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336788 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336798 4888 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336823 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336836 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336847 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336858 4888 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336897 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336908 4888 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336918 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336928 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336939 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336948 4888 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336958 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336967 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336976 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336985 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.336995 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337004 4888 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337013 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337022 4888 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337031 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337042 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337053 4888 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337063 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337073 4888 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337082 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337091 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337101 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337112 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337121 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337131 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337141 4888 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337151 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337159 4888 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337169 4888 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337178 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337189 4888 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337199 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337209 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337218 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337228 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337237 4888 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337247 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337257 4888 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337266 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337276 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337286 4888 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337295 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337304 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337315 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337324 4888 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337333 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337342 4888 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337351 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337361 4888 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337370 4888 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337379 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337389 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337397 4888 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337407 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337415 4888 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337427 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337436 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337445 4888 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337454 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337464 4888 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337473 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337483 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337492 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337501 4888 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337510 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337519 4888 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337528 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337537 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337546 4888 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337556 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337566 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337576 4888 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337588 4888 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337598 4888 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337609 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337618 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337627 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337636 4888 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337648 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337658 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337667 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337677 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337686 4888 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337695 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337704 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337713 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337723 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337732 4888 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337741 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337752 4888 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337761 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337770 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337780 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337789 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337799 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337822 4888 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337833 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337842 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337852 4888 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337864 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337873 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337882 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337891 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337900 4888 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337909 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337918 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337927 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337936 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337945 4888 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337954 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337962 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337972 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337982 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.337991 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338000 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338009 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338018 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338029 4888 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338038 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338047 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338056 4888 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338066 4888 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338077 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338087 4888 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338096 4888 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338107 4888 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338116 4888 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338125 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338135 4888 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338145 4888 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338154 4888 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338163 4888 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338173 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338183 4888 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338192 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338201 4888 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338211 4888 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338220 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338229 4888 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338239 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338250 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338259 4888 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338269 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338280 4888 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338289 4888 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338298 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338309 4888 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338320 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338330 4888 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338339 4888 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338347 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338356 4888 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338364 4888 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338372 4888 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338381 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.338390 4888 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.395669 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.458172 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.471388 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.490331 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.504977 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.512799 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.513443 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.554217 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.585260 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.603682 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.623483 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.641716 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.745487 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.745754 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.745973 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:18.745923726 +0000 UTC m=+21.328607790 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.846324 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.846599 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:18.84656254 +0000 UTC m=+21.429246584 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.846867 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.846936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:17 crc kubenswrapper[4888]: I1124 00:25:17.846988 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847120 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847164 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847185 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847191 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:18.847183527 +0000 UTC m=+21.429867561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847199 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847264 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:18.847254329 +0000 UTC m=+21.429938373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847270 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847301 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847337 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:17 crc kubenswrapper[4888]: E1124 00:25:17.847412 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:18.847393953 +0000 UTC m=+21.430078017 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.073292 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hwxtq"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.073721 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dpn9j"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.074006 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.074473 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.079540 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.079954 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.080175 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.080335 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.084496 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.085397 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.085427 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.114070 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.134863 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.150043 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a234307-6086-494b-9373-19b006699223-host\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.150082 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7a234307-6086-494b-9373-19b006699223-serviceca\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.150100 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r978x\" (UniqueName: \"kubernetes.io/projected/7a234307-6086-494b-9373-19b006699223-kube-api-access-r978x\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.150140 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-hosts-file\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.150158 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl7gv\" (UniqueName: \"kubernetes.io/projected/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-kube-api-access-tl7gv\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.166960 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.189867 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.218552 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.233228 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.244390 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.244557 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.247121 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.249793 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.250676 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251566 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r978x\" (UniqueName: \"kubernetes.io/projected/7a234307-6086-494b-9373-19b006699223-kube-api-access-r978x\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251632 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a234307-6086-494b-9373-19b006699223-host\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251666 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7a234307-6086-494b-9373-19b006699223-serviceca\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251745 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-hosts-file\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251791 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl7gv\" (UniqueName: \"kubernetes.io/projected/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-kube-api-access-tl7gv\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a234307-6086-494b-9373-19b006699223-host\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.251903 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-hosts-file\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.252855 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.252992 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7a234307-6086-494b-9373-19b006699223-serviceca\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.253989 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.255232 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.264432 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.265397 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.265990 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.267286 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.267895 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.268882 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.269687 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.270735 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.271553 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.276028 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r978x\" (UniqueName: \"kubernetes.io/projected/7a234307-6086-494b-9373-19b006699223-kube-api-access-r978x\") pod \"node-ca-dpn9j\" (UID: \"7a234307-6086-494b-9373-19b006699223\") " pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.276410 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.277059 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.278007 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.278586 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.279706 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.280327 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.280945 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.281423 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.281997 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.282446 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.284461 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.284902 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.286252 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.286913 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.289175 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.289762 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.290377 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.291234 4888 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.291341 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.293054 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.293991 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.294506 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.296095 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.297100 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.297739 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.298753 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.299509 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.300149 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.302638 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl7gv\" (UniqueName: \"kubernetes.io/projected/8f8109bc-1f84-4f79-96e3-a74cf1b1041c-kube-api-access-tl7gv\") pod \"node-resolver-hwxtq\" (UID: \"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\") " pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.306354 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.307184 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.308257 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.308898 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.309782 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.310357 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.311323 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.312196 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.313105 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.313412 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.313636 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.314133 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.315095 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.315686 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.316646 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.324246 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.337663 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.353664 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.366108 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.375581 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.379978 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b2d5d5114b50f8104a293337d5a4998fe865dbbf065e32d7c817699ee1a19348"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.384640 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.384723 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.384741 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1e2a14686b51387a450d33373ed79e891b667c2b965eb82ac8040672937d6e3f"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.385110 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dpn9j" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.387703 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.387748 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"661ea3f9a53078e44f76c97042b4b18c7119cc549224dfc2894a9cde162a0b6b"} Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.388365 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.392322 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hwxtq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.400430 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: W1124 00:25:18.407841 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a234307_6086_494b_9373_19b006699223.slice/crio-b2a947408b8adc85316428e8c6e1c7da54b41ec62f22062551e20c1e6f269da3 WatchSource:0}: Error finding container b2a947408b8adc85316428e8c6e1c7da54b41ec62f22062551e20c1e6f269da3: Status 404 returned error can't find the container with id b2a947408b8adc85316428e8c6e1c7da54b41ec62f22062551e20c1e6f269da3 Nov 24 00:25:18 crc kubenswrapper[4888]: W1124 00:25:18.410088 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f8109bc_1f84_4f79_96e3_a74cf1b1041c.slice/crio-4d522fc3d56904c511ff29cd5c50dae8288907bb4a89ef677450bcbdb462c9e0 WatchSource:0}: Error finding container 4d522fc3d56904c511ff29cd5c50dae8288907bb4a89ef677450bcbdb462c9e0: Status 404 returned error can't find the container with id 4d522fc3d56904c511ff29cd5c50dae8288907bb4a89ef677450bcbdb462c9e0 Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.416287 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.427777 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.438659 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.465228 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.471997 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-wvkcz"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.472496 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.474514 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.474608 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.474953 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.475084 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.475141 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.476448 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.486381 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.500882 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.517025 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.532717 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.536229 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.541170 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.550403 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.554419 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d17c94b3-864e-4594-8ed7-06819e60fa5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.554458 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d17c94b3-864e-4594-8ed7-06819e60fa5e-proxy-tls\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.554503 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d17c94b3-864e-4594-8ed7-06819e60fa5e-rootfs\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.554560 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tkt2\" (UniqueName: \"kubernetes.io/projected/d17c94b3-864e-4594-8ed7-06819e60fa5e-kube-api-access-6tkt2\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.563886 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.574632 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.587399 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.592997 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.598178 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.611642 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.625213 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.638145 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.651625 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.655861 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d17c94b3-864e-4594-8ed7-06819e60fa5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.655902 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d17c94b3-864e-4594-8ed7-06819e60fa5e-proxy-tls\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.655919 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d17c94b3-864e-4594-8ed7-06819e60fa5e-rootfs\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.655944 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tkt2\" (UniqueName: \"kubernetes.io/projected/d17c94b3-864e-4594-8ed7-06819e60fa5e-kube-api-access-6tkt2\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.656919 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d17c94b3-864e-4594-8ed7-06819e60fa5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.657347 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d17c94b3-864e-4594-8ed7-06819e60fa5e-rootfs\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.664394 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d17c94b3-864e-4594-8ed7-06819e60fa5e-proxy-tls\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.667351 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.678335 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tkt2\" (UniqueName: \"kubernetes.io/projected/d17c94b3-864e-4594-8ed7-06819e60fa5e-kube-api-access-6tkt2\") pod \"machine-config-daemon-wvkcz\" (UID: \"d17c94b3-864e-4594-8ed7-06819e60fa5e\") " pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.686710 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.756990 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.757140 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.757201 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:20.757186277 +0000 UTC m=+23.339870321 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.787679 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:25:18 crc kubenswrapper[4888]: W1124 00:25:18.799641 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd17c94b3_864e_4594_8ed7_06819e60fa5e.slice/crio-7ce61938cdcfc7ed2946c6e9875ea6a218d19f3ff34a1d03b2db0f3fd8e7db4a WatchSource:0}: Error finding container 7ce61938cdcfc7ed2946c6e9875ea6a218d19f3ff34a1d03b2db0f3fd8e7db4a: Status 404 returned error can't find the container with id 7ce61938cdcfc7ed2946c6e9875ea6a218d19f3ff34a1d03b2db0f3fd8e7db4a Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.844197 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-zbfxr"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.844616 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ltl44"] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.844770 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.845271 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.849965 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.850457 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.850517 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.850975 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.851623 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.851670 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.851675 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.857785 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.857934 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.857987 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.858016 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858092 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858143 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:20.858129909 +0000 UTC m=+23.440813953 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858444 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:20.858436407 +0000 UTC m=+23.441120451 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858531 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858546 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858557 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858584 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:20.858575581 +0000 UTC m=+23.441259625 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858626 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858637 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858645 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:18 crc kubenswrapper[4888]: E1124 00:25:18.858670 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:20.858663713 +0000 UTC m=+23.441347757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.866375 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.899180 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.917924 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.949375 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958603 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-system-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958656 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cnibin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958678 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-k8s-cni-cncf-io\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958720 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cni-binary-copy\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958762 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-os-release\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958780 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-hostroot\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958802 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-os-release\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958852 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958872 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-etc-kubernetes\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958903 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-multus-certs\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958926 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-binary-copy\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-netns\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958979 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-daemon-config\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.958999 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-system-cni-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959029 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-socket-dir-parent\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959047 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-kubelet\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959070 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-cnibin\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959105 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-multus\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959124 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-bin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959143 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ghmx\" (UniqueName: \"kubernetes.io/projected/4ba4b666-4513-46b8-b0f8-f39f56fb558f-kube-api-access-6ghmx\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959250 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-conf-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959295 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.959322 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xns87\" (UniqueName: \"kubernetes.io/projected/adcded18-1bca-43c9-ac95-40ba409b3549-kube-api-access-xns87\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.968240 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.986865 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:18 crc kubenswrapper[4888]: I1124 00:25:18.997000 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.010145 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.024529 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.043031 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.058760 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060028 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060060 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-etc-kubernetes\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060082 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-binary-copy\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060106 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-multus-certs\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060142 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-netns\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060168 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-daemon-config\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060185 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-system-cni-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060180 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-etc-kubernetes\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060204 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-socket-dir-parent\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060221 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-kubelet\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060247 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-cnibin\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060268 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-multus\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060285 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-bin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060300 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ghmx\" (UniqueName: \"kubernetes.io/projected/4ba4b666-4513-46b8-b0f8-f39f56fb558f-kube-api-access-6ghmx\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060323 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-conf-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060360 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xns87\" (UniqueName: \"kubernetes.io/projected/adcded18-1bca-43c9-ac95-40ba409b3549-kube-api-access-xns87\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060385 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-system-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060402 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-k8s-cni-cncf-io\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cnibin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060446 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cni-binary-copy\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060468 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060503 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-os-release\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060529 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-hostroot\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060551 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-os-release\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060539 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060907 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-binary-copy\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060942 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-daemon-config\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060970 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-system-cni-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.060985 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-k8s-cni-cncf-io\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061022 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cnibin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061288 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/adcded18-1bca-43c9-ac95-40ba409b3549-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061360 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-os-release\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061372 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-system-cni-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061413 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-socket-dir-parent\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061451 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-multus\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061480 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-kubelet\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061503 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-cnibin\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061530 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-os-release\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061531 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-multus-certs\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061564 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ba4b666-4513-46b8-b0f8-f39f56fb558f-cni-binary-copy\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061581 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-multus-conf-dir\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061420 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-hostroot\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061604 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-run-netns\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061624 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ba4b666-4513-46b8-b0f8-f39f56fb558f-host-var-lib-cni-bin\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.061794 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/adcded18-1bca-43c9-ac95-40ba409b3549-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.075169 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.079918 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ghmx\" (UniqueName: \"kubernetes.io/projected/4ba4b666-4513-46b8-b0f8-f39f56fb558f-kube-api-access-6ghmx\") pod \"multus-zbfxr\" (UID: \"4ba4b666-4513-46b8-b0f8-f39f56fb558f\") " pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.079924 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xns87\" (UniqueName: \"kubernetes.io/projected/adcded18-1bca-43c9-ac95-40ba409b3549-kube-api-access-xns87\") pod \"multus-additional-cni-plugins-ltl44\" (UID: \"adcded18-1bca-43c9-ac95-40ba409b3549\") " pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.091564 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.108202 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.123589 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.142480 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.158017 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ltl44" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.161411 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.165258 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zbfxr" Nov 24 00:25:19 crc kubenswrapper[4888]: W1124 00:25:19.172700 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadcded18_1bca_43c9_ac95_40ba409b3549.slice/crio-ceac1f5f486a9b59244183514b7297735ea6ddc169b147ce9b743429126a594d WatchSource:0}: Error finding container ceac1f5f486a9b59244183514b7297735ea6ddc169b147ce9b743429126a594d: Status 404 returned error can't find the container with id ceac1f5f486a9b59244183514b7297735ea6ddc169b147ce9b743429126a594d Nov 24 00:25:19 crc kubenswrapper[4888]: W1124 00:25:19.178802 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ba4b666_4513_46b8_b0f8_f39f56fb558f.slice/crio-4bdb4b69936a34b893c22a450dab71e4b47ce5cc3148564923201c369086171c WatchSource:0}: Error finding container 4bdb4b69936a34b893c22a450dab71e4b47ce5cc3148564923201c369086171c: Status 404 returned error can't find the container with id 4bdb4b69936a34b893c22a450dab71e4b47ce5cc3148564923201c369086171c Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.179026 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.200665 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.215934 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.231881 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.244334 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.244358 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:19 crc kubenswrapper[4888]: E1124 00:25:19.244479 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:19 crc kubenswrapper[4888]: E1124 00:25:19.244702 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.254609 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.258482 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6j5lq"] Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.259656 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.266002 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.266282 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.266358 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.266497 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.266602 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.269493 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.269723 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.282375 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.298923 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.315382 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.335107 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.352320 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.363847 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364114 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364162 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364274 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364311 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364340 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364373 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364478 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364539 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364574 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364627 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364655 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364711 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz76r\" (UniqueName: \"kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364734 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364761 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364844 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364874 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364900 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364933 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364978 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.364997 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.378542 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.393155 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.393228 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.393252 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"7ce61938cdcfc7ed2946c6e9875ea6a218d19f3ff34a1d03b2db0f3fd8e7db4a"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.395156 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hwxtq" event={"ID":"8f8109bc-1f84-4f79-96e3-a74cf1b1041c","Type":"ContainerStarted","Data":"6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.395235 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hwxtq" event={"ID":"8f8109bc-1f84-4f79-96e3-a74cf1b1041c","Type":"ContainerStarted","Data":"4d522fc3d56904c511ff29cd5c50dae8288907bb4a89ef677450bcbdb462c9e0"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.397021 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerStarted","Data":"2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.397048 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerStarted","Data":"4bdb4b69936a34b893c22a450dab71e4b47ce5cc3148564923201c369086171c"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.398205 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerStarted","Data":"ceac1f5f486a9b59244183514b7297735ea6ddc169b147ce9b743429126a594d"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.400492 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dpn9j" event={"ID":"7a234307-6086-494b-9373-19b006699223","Type":"ContainerStarted","Data":"777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.400537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dpn9j" event={"ID":"7a234307-6086-494b-9373-19b006699223","Type":"ContainerStarted","Data":"b2a947408b8adc85316428e8c6e1c7da54b41ec62f22062551e20c1e6f269da3"} Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.407232 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.449069 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466084 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466126 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466154 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466171 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466245 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz76r\" (UniqueName: \"kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466271 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466289 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466328 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466355 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466369 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466383 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466398 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466443 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466492 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466510 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466533 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466547 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.466563 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475105 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475199 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475237 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475316 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475663 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.475999 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476052 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476061 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476093 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476091 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476113 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476123 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476158 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476538 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476761 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.476974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.477020 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.479032 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.482415 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.490709 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.517357 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz76r\" (UniqueName: \"kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r\") pod \"ovnkube-node-6j5lq\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.546881 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.590149 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.595675 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:19 crc kubenswrapper[4888]: W1124 00:25:19.609689 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d662e59_55e9_45fa_870d_ad8c51a7a0ef.slice/crio-f48f416a055c9523d571a98976942868f588e6d93f4c1858e9aede4a05dc8947 WatchSource:0}: Error finding container f48f416a055c9523d571a98976942868f588e6d93f4c1858e9aede4a05dc8947: Status 404 returned error can't find the container with id f48f416a055c9523d571a98976942868f588e6d93f4c1858e9aede4a05dc8947 Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.640991 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.667719 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.708276 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.746933 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.790004 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.826307 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.869468 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.906484 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:19 crc kubenswrapper[4888]: I1124 00:25:19.978091 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.015390 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.039716 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.073307 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.114416 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.148057 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.194770 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.225834 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.244776 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.244958 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.267160 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.307056 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.345600 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.405797 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732" exitCode=0 Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.405901 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732"} Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.413643 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" exitCode=0 Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.413753 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.413849 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"f48f416a055c9523d571a98976942868f588e6d93f4c1858e9aede4a05dc8947"} Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.423366 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a"} Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.431257 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.446648 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.467006 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.506203 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.544052 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.584358 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.629363 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.665265 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.743616 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.760656 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.780549 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.780749 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.780845 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:24.780826677 +0000 UTC m=+27.363510731 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.787518 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.829273 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.868082 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.881368 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.881628 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:24.881582854 +0000 UTC m=+27.464266908 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.881750 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.881860 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.881996 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882014 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882046 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882064 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882132 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:24.882113009 +0000 UTC m=+27.464797063 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882158 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882243 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:24.882211921 +0000 UTC m=+27.464896045 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882379 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882401 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882415 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:20 crc kubenswrapper[4888]: E1124 00:25:20.882454 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:24.882443768 +0000 UTC m=+27.465127812 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.908945 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.945875 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:20 crc kubenswrapper[4888]: I1124 00:25:20.988306 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.026119 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.064677 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.108379 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.148196 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.186826 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.224695 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.244960 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.245001 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:21 crc kubenswrapper[4888]: E1124 00:25:21.245136 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:21 crc kubenswrapper[4888]: E1124 00:25:21.245331 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.266830 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.313087 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.347433 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.387140 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.429012 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437253 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437360 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437392 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437422 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437449 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.437479 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.440079 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67" exitCode=0 Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.440365 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67"} Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.477324 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.510693 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.552922 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.600081 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.644855 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.669032 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.706238 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.749217 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.789418 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.826526 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.866728 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.911139 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.947722 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.965063 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.980058 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 00:25:21 crc kubenswrapper[4888]: I1124 00:25:21.983632 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:21Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.009446 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.042447 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.083877 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.123719 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.167279 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.210296 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.244280 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:22 crc kubenswrapper[4888]: E1124 00:25:22.244439 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.252285 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.294051 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.334980 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.375323 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.405754 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.448341 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a" exitCode=0 Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.448387 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a"} Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.451896 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.489282 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.528472 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.576552 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.606631 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.651235 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.687997 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.725718 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.770568 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.810209 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.848890 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.897479 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.931254 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:22 crc kubenswrapper[4888]: I1124 00:25:22.971096 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:22Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.010739 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.047712 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.084877 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.128951 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.172606 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.214258 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.245085 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.245089 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.245321 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.245439 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.262989 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.415514 4888 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.418627 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.418694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.418715 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.418859 4888 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.428464 4888 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.428900 4888 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.430577 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.430647 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.430663 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.430685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.430700 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.445422 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.453258 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.453285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.453295 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.453308 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.453317 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.467387 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09" exitCode=0 Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.467488 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09"} Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.479316 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.485079 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.485127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.485137 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.485155 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.485170 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.494635 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.501714 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.506183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.506226 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.506240 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.506265 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.506282 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.512372 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.525400 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.525902 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.531206 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.531255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.531270 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.531291 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.531304 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.543941 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.549032 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: E1124 00:25:23.549380 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.551401 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.551428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.551439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.551458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.551472 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.568513 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.585726 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.603230 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.624197 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.647552 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.654608 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.654654 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.654670 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.654691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.654705 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.686928 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.727753 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.761601 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.761671 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.761691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.761724 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.761748 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.766772 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.806720 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.853366 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.864500 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.864541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.864555 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.864575 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.864594 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.891486 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:23Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.968024 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.968110 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.968132 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.968164 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:23 crc kubenswrapper[4888]: I1124 00:25:23.968187 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:23Z","lastTransitionTime":"2025-11-24T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.071449 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.071495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.071510 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.071533 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.071547 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.175296 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.175392 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.175416 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.175907 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.176598 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.245082 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.245589 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.293333 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.293414 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.293436 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.293465 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.293483 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.398374 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.398463 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.398489 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.398516 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.398539 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.487046 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d" exitCode=0 Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.487476 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.498229 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.502878 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.502922 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.502933 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.502951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.502964 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.514089 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.548244 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.570136 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.587058 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.604896 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.611417 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.611476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.611510 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.611541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.611560 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.627306 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.652378 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.671743 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.693235 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.710557 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.716402 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.716445 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.716455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.716469 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.716480 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.726508 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.739442 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.754235 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.769264 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.784623 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:24Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.819167 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.819210 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.819219 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.819234 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.819245 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.832046 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.832258 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.832371 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:32.832347614 +0000 UTC m=+35.415031658 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.923094 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.923155 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.923171 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.923197 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.923216 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:24Z","lastTransitionTime":"2025-11-24T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.932616 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.932785 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:32.932751852 +0000 UTC m=+35.515435936 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.932918 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.933031 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933077 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933162 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:32.933140942 +0000 UTC m=+35.515824976 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933215 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933238 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933256 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933355 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:32.933295217 +0000 UTC m=+35.515979291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933443 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933470 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933486 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:24 crc kubenswrapper[4888]: E1124 00:25:24.933530 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:32.933517263 +0000 UTC m=+35.516201347 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:24 crc kubenswrapper[4888]: I1124 00:25:24.933078 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.027245 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.027301 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.027314 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.027334 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.027349 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.131358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.131400 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.131409 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.131429 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.131444 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.235280 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.235334 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.235347 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.235373 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.235390 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.244514 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:25 crc kubenswrapper[4888]: E1124 00:25:25.244712 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.245196 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:25 crc kubenswrapper[4888]: E1124 00:25:25.245352 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.339367 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.339467 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.339495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.339534 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.339560 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.443182 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.443250 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.443270 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.443299 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.443319 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.515271 4888 generic.go:334] "Generic (PLEG): container finished" podID="adcded18-1bca-43c9-ac95-40ba409b3549" containerID="a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8" exitCode=0 Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.515351 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerDied","Data":"a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.539881 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.545941 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.546011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.546027 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.546050 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.546064 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.558462 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.573270 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.589560 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.623708 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.646299 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.651694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.651727 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.651736 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.651753 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.651765 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.665667 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.684129 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.699941 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.716469 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.735237 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.754653 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.757536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.757562 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.757571 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.757589 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.757601 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.795282 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.812786 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.834685 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:25Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.860886 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.860937 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.860951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.860972 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.860987 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.964141 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.964206 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.964221 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.964239 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:25 crc kubenswrapper[4888]: I1124 00:25:25.964252 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:25Z","lastTransitionTime":"2025-11-24T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.067871 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.068393 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.068407 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.068427 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.068449 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.171161 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.171507 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.171662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.171683 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.171704 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.244615 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:26 crc kubenswrapper[4888]: E1124 00:25:26.244924 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.274150 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.274231 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.274280 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.274305 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.274322 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.377587 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.377690 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.377729 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.377759 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.377786 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.481280 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.481417 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.481439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.481485 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.481509 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.529097 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.529908 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.546480 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" event={"ID":"adcded18-1bca-43c9-ac95-40ba409b3549","Type":"ContainerStarted","Data":"9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.560687 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.569336 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.580118 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.586048 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.586112 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.586129 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.586154 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.586166 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.591852 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.605949 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.621589 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.635917 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.650975 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.683893 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.689869 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.689934 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.689951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.689977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.689993 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.703355 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.718617 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.734188 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.751131 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.769078 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.792984 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.793053 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.793074 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.793120 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.793138 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.793858 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.808350 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.827583 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.843418 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.856020 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.869246 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.894109 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.896074 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.896140 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.896171 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.896188 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.896201 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.911761 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.931860 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.949687 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.963602 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.981035 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.997519 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:26Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.999246 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.999289 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.999301 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.999320 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:26 crc kubenswrapper[4888]: I1124 00:25:26.999332 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:26Z","lastTransitionTime":"2025-11-24T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.017166 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.036293 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.052165 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.071694 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.102104 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.102152 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.102162 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.102180 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.102192 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.205389 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.205441 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.205454 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.205473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.205485 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.244446 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:27 crc kubenswrapper[4888]: E1124 00:25:27.244683 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.245158 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:27 crc kubenswrapper[4888]: E1124 00:25:27.245412 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.308420 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.308476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.308487 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.308505 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.308515 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.412151 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.412211 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.412224 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.412247 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.412267 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.515951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.516055 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.516084 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.516126 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.516154 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.550411 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.550464 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.583401 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.603919 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.619834 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.619893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.619911 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.619937 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.619955 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.625364 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.640239 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.656504 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.682347 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.699307 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.717395 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.723154 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.723210 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.723225 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.723251 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.723265 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.730403 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.744002 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.757458 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.774918 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.789564 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.803186 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.824094 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.827091 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.827158 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.827184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.827214 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.827243 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.854754 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:27Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.930635 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.930698 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.930715 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.930736 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:27 crc kubenswrapper[4888]: I1124 00:25:27.930756 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:27Z","lastTransitionTime":"2025-11-24T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.037217 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.037267 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.037277 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.037297 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.037311 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.141127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.141177 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.141189 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.141206 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.141215 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.243952 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.243994 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.244009 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.244028 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.244041 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.244139 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:28 crc kubenswrapper[4888]: E1124 00:25:28.244250 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.259014 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.277120 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.294066 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.311371 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.333886 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.345767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.345831 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.345845 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.345866 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.345880 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.353750 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.374660 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.391705 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.405578 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.420523 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.435429 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.447727 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.447760 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.447770 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.447784 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.447794 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.448524 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.461510 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.477305 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.499883 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:28Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.550303 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.550358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.550371 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.550395 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.550410 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.653495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.653543 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.653551 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.653567 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.653580 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.756720 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.756847 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.756865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.756893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.756910 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.874072 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.874136 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.874155 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.874182 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.874203 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.978219 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.978641 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.978839 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.979001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:28 crc kubenswrapper[4888]: I1124 00:25:28.979165 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:28Z","lastTransitionTime":"2025-11-24T00:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.082868 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.083279 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.083421 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.083569 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.083774 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.187529 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.188545 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.188615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.188658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.188686 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.244992 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.245023 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:29 crc kubenswrapper[4888]: E1124 00:25:29.245261 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:29 crc kubenswrapper[4888]: E1124 00:25:29.245341 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.291928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.291989 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.292003 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.292024 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.292038 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.396579 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.396660 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.396680 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.396709 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.396728 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.500222 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.500283 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.500303 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.500330 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.500350 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.559643 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/0.log" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.563442 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2" exitCode=1 Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.563490 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.564779 4888 scope.go:117] "RemoveContainer" containerID="085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.605100 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.611116 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.611165 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.611179 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.611201 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.611215 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.631989 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.658547 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.677092 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.695582 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.715778 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.715831 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.715845 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.715866 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.715880 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.717357 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.738148 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.753911 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.769538 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.795095 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.818419 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.818475 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.818489 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.818510 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.818522 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.822981 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:28Z\\\",\\\"message\\\":\\\"r removal\\\\nI1124 00:25:28.821466 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 00:25:28.821478 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 00:25:28.821492 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 00:25:28.821575 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 00:25:28.821583 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 00:25:28.821596 6193 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 00:25:28.821621 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 00:25:28.821623 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 00:25:28.821659 6193 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 00:25:28.821682 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 00:25:28.821716 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 00:25:28.821732 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 00:25:28.821755 6193 factory.go:656] Stopping watch factory\\\\nI1124 00:25:28.821769 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 00:25:28.821777 6193 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:25:28.821775 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.842132 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.857950 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.874658 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.892551 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:29Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.921798 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.921874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.921883 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.921902 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:29 crc kubenswrapper[4888]: I1124 00:25:29.921913 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:29Z","lastTransitionTime":"2025-11-24T00:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.025418 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.025521 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.025543 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.025581 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.025602 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.128882 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.128930 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.128942 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.128961 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.128976 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.231963 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.232015 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.232027 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.232045 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.232058 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.244303 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:30 crc kubenswrapper[4888]: E1124 00:25:30.244440 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.334996 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.335049 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.335067 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.335091 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.335111 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.438220 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.438280 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.438293 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.438318 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.438333 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.541249 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.541293 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.541301 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.541317 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.541329 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.568991 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/0.log" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.573133 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.573591 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.593541 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.618848 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.640126 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.644704 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.644758 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.644767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.644799 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.644825 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.665913 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.695268 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.730696 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.748165 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.748251 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.748282 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.748319 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.748344 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.749029 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.768078 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.782872 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.800350 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.826198 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.843883 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.851876 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.851959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.851981 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.852011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.852032 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.866607 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.886955 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.916609 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:28Z\\\",\\\"message\\\":\\\"r removal\\\\nI1124 00:25:28.821466 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 00:25:28.821478 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 00:25:28.821492 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 00:25:28.821575 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 00:25:28.821583 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 00:25:28.821596 6193 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 00:25:28.821621 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 00:25:28.821623 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 00:25:28.821659 6193 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 00:25:28.821682 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 00:25:28.821716 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 00:25:28.821732 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 00:25:28.821755 6193 factory.go:656] Stopping watch factory\\\\nI1124 00:25:28.821769 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 00:25:28.821777 6193 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:25:28.821775 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:30Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.955764 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.955925 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.955945 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.955972 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:30 crc kubenswrapper[4888]: I1124 00:25:30.956021 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:30Z","lastTransitionTime":"2025-11-24T00:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.059254 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.059351 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.059365 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.059389 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.059404 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.163054 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.163126 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.163145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.163173 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.163191 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.244166 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.244192 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:31 crc kubenswrapper[4888]: E1124 00:25:31.244382 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:31 crc kubenswrapper[4888]: E1124 00:25:31.244550 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.266725 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.266802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.266860 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.266896 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.266920 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.370730 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.370927 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.371001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.371080 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.371105 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.475890 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.475943 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.475955 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.475974 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.475988 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.579009 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.579082 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.579098 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.579119 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.579136 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.580947 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/1.log" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.582096 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/0.log" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.585772 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a" exitCode=1 Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.585846 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.585942 4888 scope.go:117] "RemoveContainer" containerID="085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.587301 4888 scope.go:117] "RemoveContainer" containerID="fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a" Nov 24 00:25:31 crc kubenswrapper[4888]: E1124 00:25:31.587668 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.613338 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.628981 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.642993 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.652864 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.666414 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.683619 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.683700 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.683724 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.683758 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.683789 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.686243 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.719102 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:28Z\\\",\\\"message\\\":\\\"r removal\\\\nI1124 00:25:28.821466 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 00:25:28.821478 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 00:25:28.821492 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 00:25:28.821575 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 00:25:28.821583 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 00:25:28.821596 6193 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 00:25:28.821621 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 00:25:28.821623 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 00:25:28.821659 6193 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 00:25:28.821682 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 00:25:28.821716 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 00:25:28.821732 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 00:25:28.821755 6193 factory.go:656] Stopping watch factory\\\\nI1124 00:25:28.821769 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 00:25:28.821777 6193 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:25:28.821775 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.738705 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.758974 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.776531 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.789941 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.790256 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.790397 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.790602 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.790749 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.805081 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.831987 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.847646 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs"] Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.848451 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.857292 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.857474 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.862170 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.894596 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.894678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.894703 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.894739 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.894766 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.923731 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/679c60ec-6115-419c-9074-da8a6e01b748-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.923864 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.923974 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.924083 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrm9\" (UniqueName: \"kubernetes.io/projected/679c60ec-6115-419c-9074-da8a6e01b748-kube-api-access-plrm9\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.952224 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.976195 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:31Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.997917 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.997979 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.997994 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.998016 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:31 crc kubenswrapper[4888]: I1124 00:25:31.998031 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:31Z","lastTransitionTime":"2025-11-24T00:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.004087 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.018924 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.025044 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.025118 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.025175 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrm9\" (UniqueName: \"kubernetes.io/projected/679c60ec-6115-419c-9074-da8a6e01b748-kube-api-access-plrm9\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.025274 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/679c60ec-6115-419c-9074-da8a6e01b748-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.025962 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.026303 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/679c60ec-6115-419c-9074-da8a6e01b748-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.038063 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/679c60ec-6115-419c-9074-da8a6e01b748-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.041505 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.050541 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrm9\" (UniqueName: \"kubernetes.io/projected/679c60ec-6115-419c-9074-da8a6e01b748-kube-api-access-plrm9\") pod \"ovnkube-control-plane-749d76644c-k6trs\" (UID: \"679c60ec-6115-419c-9074-da8a6e01b748\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.057599 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.072123 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.088876 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.101249 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.101311 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.101321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.101340 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.101368 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.110099 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.129976 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.147784 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.164079 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.189279 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.189984 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:28Z\\\",\\\"message\\\":\\\"r removal\\\\nI1124 00:25:28.821466 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 00:25:28.821478 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 00:25:28.821492 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 00:25:28.821575 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 00:25:28.821583 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 00:25:28.821596 6193 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 00:25:28.821621 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 00:25:28.821623 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 00:25:28.821659 6193 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 00:25:28.821682 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 00:25:28.821716 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 00:25:28.821732 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 00:25:28.821755 6193 factory.go:656] Stopping watch factory\\\\nI1124 00:25:28.821769 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 00:25:28.821777 6193 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:25:28.821775 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.204803 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.205052 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.205198 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.205344 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.205508 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.208892 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: W1124 00:25:32.212153 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod679c60ec_6115_419c_9074_da8a6e01b748.slice/crio-a49a1a2117810c39c12b61965706b71f77b5e7657f05f326b97bf41fd62907d3 WatchSource:0}: Error finding container a49a1a2117810c39c12b61965706b71f77b5e7657f05f326b97bf41fd62907d3: Status 404 returned error can't find the container with id a49a1a2117810c39c12b61965706b71f77b5e7657f05f326b97bf41fd62907d3 Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.225652 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.246974 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.247219 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.253397 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.273114 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.292047 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.308425 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.308461 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.308471 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.308492 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.308505 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.411470 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.411503 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.411511 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.411525 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.411534 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.514920 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.514980 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.514992 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.515011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.515024 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.591958 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" event={"ID":"679c60ec-6115-419c-9074-da8a6e01b748","Type":"ContainerStarted","Data":"d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.592024 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" event={"ID":"679c60ec-6115-419c-9074-da8a6e01b748","Type":"ContainerStarted","Data":"7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.592037 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" event={"ID":"679c60ec-6115-419c-9074-da8a6e01b748","Type":"ContainerStarted","Data":"a49a1a2117810c39c12b61965706b71f77b5e7657f05f326b97bf41fd62907d3"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.596146 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/1.log" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.612193 4888 scope.go:117] "RemoveContainer" containerID="fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a" Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.612476 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.617783 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.618127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.618163 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.618176 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.618195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.618210 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.636691 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.653951 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.678465 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.706013 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.721352 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.721417 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.721433 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.721456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.721471 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.729923 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.748651 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.763593 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.776637 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.789190 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.804560 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.816734 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.824984 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.825198 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.825331 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.825481 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.825614 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.829995 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.836690 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.837000 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.837105 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.83707782 +0000 UTC m=+51.419761874 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.847044 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.874622 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://085a896fb7f9e57be258b51e9c3e9137e1e25567474f2a92bf2c83af02ffead2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:28Z\\\",\\\"message\\\":\\\"r removal\\\\nI1124 00:25:28.821466 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 00:25:28.821478 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 00:25:28.821492 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 00:25:28.821575 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 00:25:28.821583 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 00:25:28.821596 6193 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 00:25:28.821621 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 00:25:28.821623 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 00:25:28.821659 6193 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 00:25:28.821682 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 00:25:28.821716 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 00:25:28.821732 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 00:25:28.821755 6193 factory.go:656] Stopping watch factory\\\\nI1124 00:25:28.821769 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 00:25:28.821777 6193 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:25:28.821775 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.888071 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.910420 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.928134 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.928189 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.928205 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.928225 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.928238 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:32Z","lastTransitionTime":"2025-11-24T00:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.931498 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.937329 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.937416 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.937447 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937480 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.937450136 +0000 UTC m=+51.520134180 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.937539 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937592 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937660 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937592 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937694 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937707 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937766 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.937745744 +0000 UTC m=+51.520429798 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937719 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937802 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937847 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.937798956 +0000 UTC m=+51.520483020 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:32 crc kubenswrapper[4888]: E1124 00:25:32.937904 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.937878398 +0000 UTC m=+51.520562452 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.949756 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.961874 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.973896 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.984122 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:32 crc kubenswrapper[4888]: I1124 00:25:32.996438 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:32Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.009636 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.012510 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7j4n7"] Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.013393 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.013496 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.025620 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.031718 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.031843 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.031865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.031893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.031914 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.038210 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whnmn\" (UniqueName: \"kubernetes.io/projected/ee556f32-31ee-45ef-bf08-26bf19390c96-kube-api-access-whnmn\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.038325 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.041053 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.063405 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.086177 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.103265 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.118793 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135344 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135454 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135641 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135885 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135932 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.135968 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.138851 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.138913 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whnmn\" (UniqueName: \"kubernetes.io/projected/ee556f32-31ee-45ef-bf08-26bf19390c96-kube-api-access-whnmn\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.139062 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.139156 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:33.639126355 +0000 UTC m=+36.221810599 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.163500 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.169678 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whnmn\" (UniqueName: \"kubernetes.io/projected/ee556f32-31ee-45ef-bf08-26bf19390c96-kube-api-access-whnmn\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.191964 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.219889 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.235657 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.239156 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.239202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.239215 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.239236 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.239251 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.244769 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.244935 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.245091 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.245324 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.253532 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.268347 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.281044 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.291583 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.305218 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.325112 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.341703 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.342125 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.342153 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.342170 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.342196 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.342217 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.355189 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.388017 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.418773 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.436712 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.445419 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.445507 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.445522 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.445541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.445554 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.453627 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.467374 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.485753 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.548667 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.548718 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.548731 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.548753 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.548769 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.644670 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.644893 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.644981 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:34.644957805 +0000 UTC m=+37.227641879 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.651955 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.652017 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.652037 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.652061 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.652080 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.756257 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.756335 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.756358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.756390 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.756412 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.860334 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.860428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.860448 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.860947 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.861367 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.903782 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.903895 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.903916 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.903943 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.903962 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.926962 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.933148 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.933233 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.933258 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.933291 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.933310 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.956145 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.962010 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.962102 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.962130 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.962165 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.962186 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:33 crc kubenswrapper[4888]: E1124 00:25:33.985057 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:33Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.990903 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.990972 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.990997 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.991029 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:33 crc kubenswrapper[4888]: I1124 00:25:33.991054 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:33Z","lastTransitionTime":"2025-11-24T00:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.012966 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:34Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.019276 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.019345 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.019371 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.019406 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.019431 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.039501 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:34Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.039769 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.042404 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.042498 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.042519 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.042546 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.042565 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.146072 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.146145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.146161 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.146184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.146200 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.244483 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.244752 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.251327 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.251523 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.251561 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.251594 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.251617 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.355788 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.355924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.355956 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.355994 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.356038 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.459577 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.459644 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.459664 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.459691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.459710 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.563157 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.563218 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.563231 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.563254 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.563268 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.657098 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.657361 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:34 crc kubenswrapper[4888]: E1124 00:25:34.657473 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:36.657439948 +0000 UTC m=+39.240124032 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.667866 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.668239 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.668398 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.668598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.668732 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.772334 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.772439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.772467 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.772506 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.772535 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.876127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.876186 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.876199 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.876222 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.876243 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.980669 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.980742 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.980767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.980801 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:34 crc kubenswrapper[4888]: I1124 00:25:34.980855 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:34Z","lastTransitionTime":"2025-11-24T00:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.083769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.083873 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.083894 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.083928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.083948 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.187017 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.188041 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.188279 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.188441 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.188585 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.244955 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.244986 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.245147 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:35 crc kubenswrapper[4888]: E1124 00:25:35.245754 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:35 crc kubenswrapper[4888]: E1124 00:25:35.245870 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:35 crc kubenswrapper[4888]: E1124 00:25:35.246075 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.292623 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.292663 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.292674 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.292691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.292703 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.397100 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.397161 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.397179 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.397207 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.397226 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.501174 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.501279 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.501307 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.501347 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.501388 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.605597 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.605652 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.605664 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.605685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.605702 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.708607 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.708657 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.708668 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.708741 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.708757 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.812979 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.813056 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.813073 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.813103 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.813122 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.916765 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.916884 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.916902 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.916927 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:35 crc kubenswrapper[4888]: I1124 00:25:35.916945 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:35Z","lastTransitionTime":"2025-11-24T00:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.020929 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.020997 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.021021 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.021050 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.021068 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.125987 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.126424 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.126567 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.126771 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.127014 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.231037 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.231161 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.231180 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.231213 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.231234 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.245583 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:36 crc kubenswrapper[4888]: E1124 00:25:36.245860 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.335624 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.335708 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.335729 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.335762 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.335784 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.439968 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.440028 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.440054 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.440090 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.440114 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.544236 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.544321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.544345 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.544378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.544403 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.648329 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.648394 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.648413 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.648440 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.648459 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.681039 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:36 crc kubenswrapper[4888]: E1124 00:25:36.681336 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:36 crc kubenswrapper[4888]: E1124 00:25:36.681490 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:40.68145548 +0000 UTC m=+43.264139554 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.752507 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.752587 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.752613 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.752646 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.752669 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.856128 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.856193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.856211 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.856238 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.856255 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.960168 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.960239 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.960258 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.960283 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:36 crc kubenswrapper[4888]: I1124 00:25:36.960301 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:36Z","lastTransitionTime":"2025-11-24T00:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.064043 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.064148 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.064169 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.064197 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.064218 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.168160 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.168252 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.168269 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.168294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.168315 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.244792 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.244935 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.244935 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:37 crc kubenswrapper[4888]: E1124 00:25:37.245112 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:37 crc kubenswrapper[4888]: E1124 00:25:37.245395 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:37 crc kubenswrapper[4888]: E1124 00:25:37.245658 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.271368 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.271459 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.271484 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.271518 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.271545 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.376318 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.376398 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.376423 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.376460 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.376486 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.480038 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.480121 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.480144 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.480176 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.480197 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.584805 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.584925 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.584949 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.584983 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.585009 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.689101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.689199 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.689230 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.689264 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.689290 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.793720 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.793854 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.793874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.793902 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.793921 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.897351 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.897434 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.897456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.897490 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:37 crc kubenswrapper[4888]: I1124 00:25:37.897513 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:37Z","lastTransitionTime":"2025-11-24T00:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.000836 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.000914 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.000939 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.000968 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.000994 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.104648 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.104732 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.104752 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.104780 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.104799 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.208116 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.208186 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.208203 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.208231 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.208252 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.244774 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:38 crc kubenswrapper[4888]: E1124 00:25:38.245090 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.264355 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.283844 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.304944 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.321051 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.321763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.321894 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.322124 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.322213 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.339223 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.362634 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.380723 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.403461 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.426011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.426076 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.426101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.426128 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.426148 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.434344 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.452571 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.474476 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.491734 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.511866 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.529612 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.529675 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.529692 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.529718 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.529735 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.532228 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.544930 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.567904 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.581019 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.603495 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:38Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.632802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.632901 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.632924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.632955 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.632977 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.736595 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.736657 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.736680 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.736712 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.736736 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.840389 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.840448 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.840464 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.840491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.840509 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.944070 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.944124 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.944144 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.944174 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:38 crc kubenswrapper[4888]: I1124 00:25:38.944194 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:38Z","lastTransitionTime":"2025-11-24T00:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.047467 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.047536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.047555 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.047583 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.047603 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.151103 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.151183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.151202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.151229 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.151249 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.245101 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.245271 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:39 crc kubenswrapper[4888]: E1124 00:25:39.245369 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:39 crc kubenswrapper[4888]: E1124 00:25:39.245495 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.245101 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:39 crc kubenswrapper[4888]: E1124 00:25:39.245634 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.255035 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.255631 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.255948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.256153 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.256293 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.368086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.368154 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.368173 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.368199 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.368217 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.471620 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.471691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.471709 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.471878 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.471911 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.575632 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.575691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.575717 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.575747 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.575771 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.679016 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.679076 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.679100 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.679135 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.679162 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.782274 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.782332 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.782352 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.782376 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.782394 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.886328 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.886473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.886541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.886569 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.886623 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.991001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.991090 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.991150 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.991177 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:39 crc kubenswrapper[4888]: I1124 00:25:39.991656 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:39Z","lastTransitionTime":"2025-11-24T00:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.096125 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.096183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.096222 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.096255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.096280 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.201709 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.201863 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.201886 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.201915 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.201935 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.244958 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:40 crc kubenswrapper[4888]: E1124 00:25:40.245215 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.304940 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.305027 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.305051 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.305080 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.305102 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.415686 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.415769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.415796 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.415860 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.415887 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.519618 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.519694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.519733 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.519763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.519786 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.623134 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.623213 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.623236 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.623270 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.623292 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.726456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.726518 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.726537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.726563 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.726583 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.731487 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:40 crc kubenswrapper[4888]: E1124 00:25:40.731739 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:40 crc kubenswrapper[4888]: E1124 00:25:40.731904 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:25:48.731869758 +0000 UTC m=+51.314553842 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.829867 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.830403 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.830598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.830769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.830990 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.934943 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.935326 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.935498 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.935653 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:40 crc kubenswrapper[4888]: I1124 00:25:40.935788 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:40Z","lastTransitionTime":"2025-11-24T00:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.039079 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.039160 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.039178 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.039207 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.039228 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.142465 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.142535 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.142559 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.142593 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.142616 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.244141 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.244221 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.244404 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:41 crc kubenswrapper[4888]: E1124 00:25:41.244641 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:41 crc kubenswrapper[4888]: E1124 00:25:41.244805 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:41 crc kubenswrapper[4888]: E1124 00:25:41.245129 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.245679 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.245794 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.245903 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.245936 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.245962 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.355398 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.355481 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.355503 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.355537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.355566 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.460152 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.460705 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.461288 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.461640 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.462002 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.565404 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.565480 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.565496 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.565515 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.565531 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.668472 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.668571 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.668599 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.668631 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.668654 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.772321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.772404 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.772428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.772458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.772482 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.876075 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.876175 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.876200 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.876235 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.876254 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.980144 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.980252 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.980284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.980338 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:41 crc kubenswrapper[4888]: I1124 00:25:41.980393 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:41Z","lastTransitionTime":"2025-11-24T00:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.084521 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.084633 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.084647 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.084669 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.084682 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.188546 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.188612 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.188639 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.188675 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.188701 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.244279 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:42 crc kubenswrapper[4888]: E1124 00:25:42.244514 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.292417 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.292481 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.292502 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.292527 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.292542 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.396018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.396146 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.396167 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.396202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.396224 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.500202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.500265 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.500283 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.500315 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.500336 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.604711 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.604784 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.604802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.604879 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.604917 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.709264 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.709376 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.709395 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.709426 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.709451 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.817193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.817295 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.817323 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.817364 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.817393 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.921476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.921538 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.921558 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.921584 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:42 crc kubenswrapper[4888]: I1124 00:25:42.921604 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:42Z","lastTransitionTime":"2025-11-24T00:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.025914 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.026043 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.026069 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.026094 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.026147 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.129879 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.129940 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.129958 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.129983 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.130003 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.233282 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.233356 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.233377 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.233406 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.233425 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.244984 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.245037 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.245136 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:43 crc kubenswrapper[4888]: E1124 00:25:43.245309 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:43 crc kubenswrapper[4888]: E1124 00:25:43.245505 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:43 crc kubenswrapper[4888]: E1124 00:25:43.245668 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.336491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.336575 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.336598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.336632 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.336653 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.440478 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.440544 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.440568 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.440602 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.440627 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.550252 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.550328 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.550353 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.550385 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.550457 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.657169 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.657243 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.657261 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.657289 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.657307 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.761087 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.761166 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.761189 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.761219 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.761242 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.865117 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.865184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.865215 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.865248 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.865269 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.968106 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.968175 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.968193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.968217 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:43 crc kubenswrapper[4888]: I1124 00:25:43.968236 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:43Z","lastTransitionTime":"2025-11-24T00:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.072347 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.072411 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.072428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.072452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.072471 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.176096 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.176174 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.176193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.176226 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.176255 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.192402 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.192487 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.192508 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.192536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.192555 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.216013 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:44Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.223037 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.223092 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.223109 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.223132 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.223150 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.245131 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.245401 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.247144 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:44Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.255205 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.255274 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.255297 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.255327 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.255349 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.277535 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:44Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.283403 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.283455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.283471 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.283497 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.283518 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.303915 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:44Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.309874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.309946 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.309971 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.310004 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.310028 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.333745 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:44Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:44 crc kubenswrapper[4888]: E1124 00:25:44.334079 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.336369 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.336439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.336464 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.336493 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.336518 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.439990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.440052 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.440071 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.440096 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.440114 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.543949 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.544004 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.544023 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.544044 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.544058 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.647706 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.647783 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.647805 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.647873 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.647900 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.752162 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.752237 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.752261 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.752292 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.752316 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.856184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.856263 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.856286 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.856317 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.856335 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.959469 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.959580 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.959607 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.959635 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:44 crc kubenswrapper[4888]: I1124 00:25:44.959659 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:44Z","lastTransitionTime":"2025-11-24T00:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.063894 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.064058 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.064090 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.064123 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.064144 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.168032 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.168105 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.168141 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.168175 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.168198 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.244759 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.244879 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.244774 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:45 crc kubenswrapper[4888]: E1124 00:25:45.245331 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:45 crc kubenswrapper[4888]: E1124 00:25:45.245449 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:45 crc kubenswrapper[4888]: E1124 00:25:45.245558 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.245659 4888 scope.go:117] "RemoveContainer" containerID="fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.272144 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.272766 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.272906 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.272939 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.272957 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.376769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.376890 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.376910 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.376940 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.376966 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.481222 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.481316 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.481345 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.481380 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.481404 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.584963 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.585033 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.585050 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.585080 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.585102 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.668038 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/1.log" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.671469 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.672197 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.688439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.688491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.688505 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.688528 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.688543 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.694694 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.722541 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.744347 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.776024 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.791465 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.791518 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.791532 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.791556 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.791573 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.800549 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.811016 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.824437 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.846030 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.859185 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.875658 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.893115 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.895920 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.895981 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.895999 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.896023 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.896038 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.911382 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.933599 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.948886 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.980233 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.999663 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.999702 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.999711 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:45 crc kubenswrapper[4888]: I1124 00:25:45.999730 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:45.999740 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:45Z","lastTransitionTime":"2025-11-24T00:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.000596 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:45Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.018139 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.102101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.102145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.102158 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.102179 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.102193 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.205053 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.205098 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.205109 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.205124 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.205134 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.244648 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:46 crc kubenswrapper[4888]: E1124 00:25:46.244843 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.307744 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.307783 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.307795 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.307823 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.307835 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.410676 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.410717 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.410726 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.410740 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.410749 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.513394 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.513435 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.513449 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.513468 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.513482 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.615442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.615496 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.615509 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.615530 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.615544 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.676920 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/2.log" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.677592 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/1.log" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.680536 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" exitCode=1 Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.680577 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.680617 4888 scope.go:117] "RemoveContainer" containerID="fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.681300 4888 scope.go:117] "RemoveContainer" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" Nov 24 00:25:46 crc kubenswrapper[4888]: E1124 00:25:46.681477 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.701245 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.718545 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.718584 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.718594 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.718609 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.718621 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.729666 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe723145952632d4e3ba5f3dc7a34ca3c05dec8eacd1e72fa0ea805c3023360a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:30Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"be9dcc9e-c16a-4962-a6d2-4adeb0b929c4\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.749559 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.773575 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.791440 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.812685 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.822677 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.822741 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.822758 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.822792 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.822834 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.837470 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.854625 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.894076 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.916214 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.926498 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.926572 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.926592 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.926616 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.926633 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:46Z","lastTransitionTime":"2025-11-24T00:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.937120 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.956686 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:46 crc kubenswrapper[4888]: I1124 00:25:46.976898 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:46Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.006158 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.025505 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.030116 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.030170 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.030184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.030205 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.030220 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.038118 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.049256 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.133923 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.133993 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.134010 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.134036 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.134057 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.238343 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.238433 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.238460 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.238495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.238518 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.245197 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.245296 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.245197 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:47 crc kubenswrapper[4888]: E1124 00:25:47.245420 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:47 crc kubenswrapper[4888]: E1124 00:25:47.245579 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:47 crc kubenswrapper[4888]: E1124 00:25:47.245984 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.342659 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.342719 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.342740 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.342770 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.342793 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.445668 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.445749 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.445775 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.445911 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.445947 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.549452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.549613 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.549642 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.549675 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.549701 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.653297 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.653372 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.653397 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.653430 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.653454 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.687271 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/2.log" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.694647 4888 scope.go:117] "RemoveContainer" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" Nov 24 00:25:47 crc kubenswrapper[4888]: E1124 00:25:47.694958 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.718651 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.741211 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.756453 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.756553 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.756581 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.756618 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.756643 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.758868 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.777287 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.799683 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.814727 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.836572 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.860789 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.860874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.860893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.860927 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.860949 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.875705 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.903404 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.928060 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.948375 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.964727 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.964848 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.964878 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.964914 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.964944 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:47Z","lastTransitionTime":"2025-11-24T00:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.968012 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:47 crc kubenswrapper[4888]: I1124 00:25:47.987683 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:47Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.014615 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.039343 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.053255 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.067255 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.067834 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.067967 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.068076 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.068165 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.068250 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.171447 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.172018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.172184 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.172410 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.172585 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.244657 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:48 crc kubenswrapper[4888]: E1124 00:25:48.244939 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.272567 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.276224 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.276318 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.276345 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.276377 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.276399 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.291457 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.309892 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.326733 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.341358 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.361852 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.378983 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.379063 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.379083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.379107 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.379124 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.392999 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.412265 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.432527 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.451985 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.468544 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.482669 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.482723 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.482755 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.482778 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.482793 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.488436 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.504779 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.525237 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.540827 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.574544 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.585768 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.585884 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.585907 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.585972 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.586001 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.601790 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:48Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.689221 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.689286 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.689305 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.689332 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.689349 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.732091 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:48 crc kubenswrapper[4888]: E1124 00:25:48.732335 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:48 crc kubenswrapper[4888]: E1124 00:25:48.732479 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:04.732438189 +0000 UTC m=+67.315122393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.793249 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.793361 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.793388 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.793424 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.793446 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.896187 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.896248 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.896267 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.896292 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.896311 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:48Z","lastTransitionTime":"2025-11-24T00:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:48 crc kubenswrapper[4888]: I1124 00:25:48.934291 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:48 crc kubenswrapper[4888]: E1124 00:25:48.934571 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:48 crc kubenswrapper[4888]: E1124 00:25:48.934693 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:20.934661683 +0000 UTC m=+83.517345757 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.000228 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.000319 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.000347 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.000379 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.000403 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.035398 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.035645 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.035756 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.035884 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:26:21.035760599 +0000 UTC m=+83.618444683 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.035754 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.035956 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:21.035934914 +0000 UTC m=+83.618618998 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.036068 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036113 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036170 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036203 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036223 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036252 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036291 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036303 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:21.036268403 +0000 UTC m=+83.618952627 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.036370 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:21.036356025 +0000 UTC m=+83.619040099 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.104534 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.104655 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.104685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.104719 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.104759 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.208770 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.208897 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.208927 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.208956 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.208974 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.244345 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.244493 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.244582 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.244766 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.244807 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:49 crc kubenswrapper[4888]: E1124 00:25:49.244942 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.312725 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.312803 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.312863 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.312893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.312914 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.416298 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.416368 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.416385 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.416411 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.416428 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.520177 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.520241 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.520252 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.520273 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.520286 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.623919 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.623993 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.624011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.624038 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.624060 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.727372 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.727418 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.727427 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.727441 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.727454 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.830637 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.830779 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.830809 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.830893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.830920 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.934718 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.934773 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.934787 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.934860 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:49 crc kubenswrapper[4888]: I1124 00:25:49.934879 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:49Z","lastTransitionTime":"2025-11-24T00:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.038506 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.038590 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.038619 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.038652 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.038672 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.142245 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.142322 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.142337 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.142360 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.142377 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.244340 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:50 crc kubenswrapper[4888]: E1124 00:25:50.244577 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.246442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.246509 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.246527 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.246557 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.246577 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.350457 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.350524 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.350546 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.350573 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.350593 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.453790 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.453873 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.453890 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.453914 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.453930 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.557284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.557373 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.557393 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.557418 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.557436 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.660190 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.660250 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.660262 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.660282 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.660298 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.762566 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.762678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.762707 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.762747 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.762785 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.866936 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.867014 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.867034 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.867062 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.867084 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.971127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.971188 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.971203 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.971227 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:50 crc kubenswrapper[4888]: I1124 00:25:50.971241 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:50Z","lastTransitionTime":"2025-11-24T00:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.074468 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.074536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.074553 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.074579 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.074596 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.177846 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.177944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.177965 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.178018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.178059 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.245049 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.245099 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:51 crc kubenswrapper[4888]: E1124 00:25:51.245342 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.245100 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:51 crc kubenswrapper[4888]: E1124 00:25:51.245480 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:51 crc kubenswrapper[4888]: E1124 00:25:51.245934 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.288299 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.288381 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.288402 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.288439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.288505 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.393158 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.393242 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.393272 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.393312 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.393339 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.496887 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.496955 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.496971 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.496995 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.497011 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.600729 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.600785 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.600794 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.600855 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.600868 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.704281 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.704356 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.704383 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.704422 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.704451 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.808676 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.808734 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.808748 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.808770 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.808785 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.912928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.913005 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.913025 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.913054 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:51 crc kubenswrapper[4888]: I1124 00:25:51.913077 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:51Z","lastTransitionTime":"2025-11-24T00:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.016254 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.016343 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.016368 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.016392 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.016412 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.120156 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.120232 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.120251 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.120277 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.120296 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.224587 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.224660 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.224684 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.224720 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.224749 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.244724 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:52 crc kubenswrapper[4888]: E1124 00:25:52.244999 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.328874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.328989 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.329047 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.329074 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.329093 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.432357 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.432420 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.432434 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.432455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.432472 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.537011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.537092 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.537115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.537149 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.537175 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.640460 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.640550 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.640577 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.640615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.640640 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.745474 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.745549 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.745579 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.745615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.745924 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.849294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.849389 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.849419 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.849462 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.849484 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.953427 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.953498 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.953515 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.953542 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:52 crc kubenswrapper[4888]: I1124 00:25:52.953562 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:52Z","lastTransitionTime":"2025-11-24T00:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.057895 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.057996 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.058039 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.058087 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.058112 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.162632 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.162729 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.162755 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.162789 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.162860 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.244301 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.244384 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.244338 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:53 crc kubenswrapper[4888]: E1124 00:25:53.244555 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:53 crc kubenswrapper[4888]: E1124 00:25:53.244774 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:53 crc kubenswrapper[4888]: E1124 00:25:53.244967 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.267705 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.267786 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.267805 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.267911 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.267968 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.371262 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.371709 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.371928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.372122 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.372358 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.475785 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.475881 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.475903 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.475928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.475945 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.579568 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.579618 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.579634 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.579658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.579677 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.683278 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.683361 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.683386 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.683412 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.683429 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.787126 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.787191 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.787208 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.787235 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.787256 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.891737 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.891831 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.891850 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.891877 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.891897 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.994757 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.994877 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.994900 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.994938 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:53 crc kubenswrapper[4888]: I1124 00:25:53.994956 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:53Z","lastTransitionTime":"2025-11-24T00:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.097860 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.097938 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.097965 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.098001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.098024 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.201285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.201355 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.201390 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.201420 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.201443 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.244936 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.245177 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.305694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.305782 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.305805 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.305871 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.305896 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.409204 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.409282 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.409300 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.409339 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.409376 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.420275 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.420374 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.420401 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.420437 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.420463 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.444271 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.450291 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.450358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.450381 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.450408 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.450427 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.471021 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.476760 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.476848 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.476875 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.476923 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.476949 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.500123 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.506189 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.506250 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.506271 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.506341 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.506361 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.530513 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.536924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.537028 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.537047 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.537074 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.537096 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.562461 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: E1124 00:25:54.562678 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.573991 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.574138 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.574161 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.574202 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.574223 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.593170 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.614465 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.618189 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.642586 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.678115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.678229 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.678335 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.678363 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.678382 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.680617 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.706627 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.729274 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.751314 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.771291 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.781843 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.781906 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.781928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.781959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.781984 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.794546 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.822075 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.858148 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.878420 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.884752 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.884850 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.884872 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.884899 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.884924 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.903582 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.925543 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.942103 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.958000 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.977501 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.988139 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.988209 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.988234 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.988267 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.988296 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:54Z","lastTransitionTime":"2025-11-24T00:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:54 crc kubenswrapper[4888]: I1124 00:25:54.996075 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:54Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.091793 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.091881 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.091898 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.091924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.091943 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.195595 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.195767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.195795 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.195852 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.195872 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.244858 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.244869 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.245026 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:55 crc kubenswrapper[4888]: E1124 00:25:55.245129 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:55 crc kubenswrapper[4888]: E1124 00:25:55.245328 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:55 crc kubenswrapper[4888]: E1124 00:25:55.245546 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.299316 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.299382 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.299403 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.299433 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.299454 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.403247 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.403321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.403349 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.403383 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.403404 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.506604 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.506670 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.506693 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.506727 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.506749 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.609181 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.609258 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.609284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.609314 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.609336 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.712666 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.712721 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.712738 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.712763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.712780 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.815676 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.815739 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.815757 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.815782 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.815799 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.918564 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.918618 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.918628 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.918648 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:55 crc kubenswrapper[4888]: I1124 00:25:55.918661 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:55Z","lastTransitionTime":"2025-11-24T00:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.022525 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.022586 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.022600 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.022626 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.022640 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.126585 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.126643 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.126658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.126679 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.126702 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.229014 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.229078 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.229101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.229151 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.229181 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.245256 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:56 crc kubenswrapper[4888]: E1124 00:25:56.245440 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.332923 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.332990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.333040 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.333067 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.333090 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.436520 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.436594 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.436621 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.436656 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.436679 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.539727 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.539804 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.539875 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.539904 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.539921 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.642731 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.642805 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.642838 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.642859 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.642875 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.744710 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.744780 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.744798 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.744865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.744883 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.847889 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.847948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.847962 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.847994 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.848010 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.951558 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.951620 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.951642 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.951674 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:56 crc kubenswrapper[4888]: I1124 00:25:56.951697 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:56Z","lastTransitionTime":"2025-11-24T00:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.054941 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.054989 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.055001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.055020 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.055033 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.158019 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.158102 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.158121 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.158148 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.158173 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.244665 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.244722 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.244840 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:57 crc kubenswrapper[4888]: E1124 00:25:57.244856 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:57 crc kubenswrapper[4888]: E1124 00:25:57.245002 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:57 crc kubenswrapper[4888]: E1124 00:25:57.245254 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.261340 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.261397 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.261415 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.261447 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.261474 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.364241 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.364306 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.364317 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.364336 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.364350 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.467794 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.467927 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.467969 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.468008 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.468035 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.571685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.571750 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.571767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.571793 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.571840 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.674598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.674670 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.674689 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.674715 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.674740 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.778142 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.778234 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.778261 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.778290 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.778313 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.882314 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.882399 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.882420 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.882451 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.882479 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.993414 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.993494 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.993510 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.993531 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:57 crc kubenswrapper[4888]: I1124 00:25:57.993546 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:57Z","lastTransitionTime":"2025-11-24T00:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.096303 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.096354 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.096372 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.096395 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.096411 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.200516 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.200570 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.200581 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.200597 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.200609 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.244461 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:25:58 crc kubenswrapper[4888]: E1124 00:25:58.244724 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.270013 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.288515 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.303473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.303555 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.303573 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.303600 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.303621 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.307174 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.322345 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.336418 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.349784 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.368773 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.387492 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.403651 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.407802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.408039 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.408173 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.408309 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.408445 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.423094 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.453624 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.469739 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.488386 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.502777 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.512534 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.512576 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.512585 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.512602 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.512613 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.523735 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.542987 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.564726 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.586245 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:25:58Z is after 2025-08-24T17:21:41Z" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.615350 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.615412 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.615438 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.615475 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.615505 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.718683 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.718754 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.718773 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.718802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.718860 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.821541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.821599 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.821608 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.821623 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.821635 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.925004 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.925049 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.925060 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.925080 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:58 crc kubenswrapper[4888]: I1124 00:25:58.925092 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:58Z","lastTransitionTime":"2025-11-24T00:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.027852 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.027922 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.027949 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.027979 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.028000 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.130908 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.131347 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.131551 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.131788 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.132039 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.235470 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.235525 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.235545 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.235572 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.235593 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.244757 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:25:59 crc kubenswrapper[4888]: E1124 00:25:59.244949 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.245061 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:25:59 crc kubenswrapper[4888]: E1124 00:25:59.245254 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.245359 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:25:59 crc kubenswrapper[4888]: E1124 00:25:59.245447 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.338405 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.339115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.339188 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.339221 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.339243 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.441945 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.442515 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.442710 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.442969 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.443164 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.547064 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.547139 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.547158 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.547195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.547227 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.650093 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.650385 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.650458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.650536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.650607 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.755944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.756035 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.756065 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.756102 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.756128 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.859694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.859767 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.859788 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.859856 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.859883 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.964339 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.964739 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.964978 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.965163 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:25:59 crc kubenswrapper[4888]: I1124 00:25:59.965293 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:25:59Z","lastTransitionTime":"2025-11-24T00:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.068610 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.068646 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.068656 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.068673 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.068683 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.173289 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.173734 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.173944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.174122 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.174274 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.245080 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:00 crc kubenswrapper[4888]: E1124 00:26:00.245270 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.277031 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.277351 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.277494 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.277642 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.277783 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.381704 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.381911 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.382036 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.382109 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.382140 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.486057 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.486127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.486140 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.486163 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.486179 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.590865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.590939 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.590959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.590989 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.591010 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.694470 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.694524 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.694536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.694556 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.694579 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.798064 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.798125 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.798143 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.798168 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.798186 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.902028 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.902114 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.902142 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.902176 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:00 crc kubenswrapper[4888]: I1124 00:26:00.902196 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:00Z","lastTransitionTime":"2025-11-24T00:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.006697 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.006763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.006783 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.006838 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.006863 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.110018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.110082 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.110098 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.110119 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.110135 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.213147 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.213229 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.213253 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.213285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.213312 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.245041 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:01 crc kubenswrapper[4888]: E1124 00:26:01.245266 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.245887 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:01 crc kubenswrapper[4888]: E1124 00:26:01.245998 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.246082 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:01 crc kubenswrapper[4888]: E1124 00:26:01.246170 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.317394 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.317467 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.317491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.317518 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.317537 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.421724 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.421798 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.421845 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.421873 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.421893 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.526079 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.526173 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.526191 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.526218 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.526236 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.629391 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.629445 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.629458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.629482 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.629497 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.732055 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.732124 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.732138 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.732168 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.732186 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.836046 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.836128 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.836147 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.836183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.836209 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.940198 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.940258 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.940279 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.940304 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:01 crc kubenswrapper[4888]: I1124 00:26:01.940322 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:01Z","lastTransitionTime":"2025-11-24T00:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.044193 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.044275 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.044293 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.044320 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.044339 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.148348 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.148423 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.148442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.148470 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.148492 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.245045 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:02 crc kubenswrapper[4888]: E1124 00:26:02.245366 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.247171 4888 scope.go:117] "RemoveContainer" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" Nov 24 00:26:02 crc kubenswrapper[4888]: E1124 00:26:02.247578 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.251586 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.251634 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.251652 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.251676 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.251694 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.356128 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.356621 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.356798 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.356981 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.357143 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.460198 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.460237 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.460244 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.460259 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.460270 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.563084 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.563164 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.563192 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.563228 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.563253 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.666870 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.666934 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.666949 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.666970 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.666983 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.769597 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.769631 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.769640 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.769654 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.769664 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.873401 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.873478 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.873494 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.873525 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.873543 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.975694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.975745 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.975758 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.975779 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:02 crc kubenswrapper[4888]: I1124 00:26:02.975792 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:02Z","lastTransitionTime":"2025-11-24T00:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.078920 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.078977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.078987 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.079009 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.079019 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.182100 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.182167 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.182188 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.182219 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.182240 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.244994 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.245097 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:03 crc kubenswrapper[4888]: E1124 00:26:03.245186 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:03 crc kubenswrapper[4888]: E1124 00:26:03.245289 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.245103 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:03 crc kubenswrapper[4888]: E1124 00:26:03.245620 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.285294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.285341 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.285352 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.285371 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.285384 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.388472 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.388509 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.388521 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.388536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.388546 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.492670 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.492749 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.492763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.492786 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.492802 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.596615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.596662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.596671 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.596688 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.596704 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.700751 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.700879 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.700906 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.700939 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.700966 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.805531 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.805602 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.805623 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.805650 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.805670 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.908266 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.908319 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.908381 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.908434 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:03 crc kubenswrapper[4888]: I1124 00:26:03.908449 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:03Z","lastTransitionTime":"2025-11-24T00:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.011078 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.011127 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.011139 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.011157 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.011168 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.114426 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.114505 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.114524 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.114549 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.114570 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.217006 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.217072 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.217086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.217107 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.217122 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.245071 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.245267 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.262783 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.320717 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.320764 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.320776 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.320794 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.320805 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.424200 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.424246 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.424255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.424281 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.424294 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.527284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.527328 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.527337 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.527355 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.527371 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.630021 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.630068 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.630080 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.630098 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.630112 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.647225 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.647285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.647295 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.647314 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.647327 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.662940 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:04Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.667018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.667072 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.667086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.667110 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.667121 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.680665 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:04Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.684928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.685011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.685026 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.685043 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.685054 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.699775 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:04Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.703541 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.703606 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.703623 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.703646 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.703665 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.720257 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:04Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.724990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.725049 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.725064 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.725086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.725102 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.737462 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.737653 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.737739 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:26:36.737719355 +0000 UTC m=+99.320403399 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.740517 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:04Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:04 crc kubenswrapper[4888]: E1124 00:26:04.740653 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.741785 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.741826 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.741836 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.741854 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.741869 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.845480 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.845546 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.845563 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.845590 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.845611 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.949647 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.949721 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.949733 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.949753 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:04 crc kubenswrapper[4888]: I1124 00:26:04.949766 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:04Z","lastTransitionTime":"2025-11-24T00:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.052625 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.052694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.052702 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.052719 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.052729 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.156421 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.156463 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.156474 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.156491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.156502 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.244426 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.244520 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.244596 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:05 crc kubenswrapper[4888]: E1124 00:26:05.244883 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:05 crc kubenswrapper[4888]: E1124 00:26:05.245044 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:05 crc kubenswrapper[4888]: E1124 00:26:05.245705 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.258857 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.258911 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.258932 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.258958 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.258978 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.361488 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.361537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.361552 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.361571 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.361583 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.464110 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.464140 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.464151 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.464169 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.464178 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.566796 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.566846 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.566856 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.566874 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.566884 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.670473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.670531 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.670550 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.670574 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.670593 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.773649 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.773722 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.773740 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.773769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.773789 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.876373 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.876449 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.876468 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.876500 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.876522 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.979037 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.979085 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.979096 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.979115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:05 crc kubenswrapper[4888]: I1124 00:26:05.979129 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:05Z","lastTransitionTime":"2025-11-24T00:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.082906 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.082977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.082987 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.083023 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.083037 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.185906 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.185953 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.185963 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.185979 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.185993 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.244884 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:06 crc kubenswrapper[4888]: E1124 00:26:06.245077 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.288670 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.288755 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.288783 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.288878 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.288908 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.391908 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.391969 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.391987 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.392009 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.392022 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.495136 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.495183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.495195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.495215 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.495229 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.598199 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.598238 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.598248 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.598266 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.598281 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.701227 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.701294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.701319 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.701351 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.701375 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.776363 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/0.log" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.776452 4888 generic.go:334] "Generic (PLEG): container finished" podID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" containerID="2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6" exitCode=1 Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.776501 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerDied","Data":"2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.777121 4888 scope.go:117] "RemoveContainer" containerID="2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.799451 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.804587 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.804671 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.804691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.804725 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.804745 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.813937 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.831267 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.851654 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.869057 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.889162 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.905027 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.907930 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.907959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.907967 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.907982 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.907991 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:06Z","lastTransitionTime":"2025-11-24T00:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.921369 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.948881 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.962184 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.977112 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:06 crc kubenswrapper[4888]: I1124 00:26:06.993164 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:06Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.005275 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.010454 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.010503 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.010515 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.010534 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.010546 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.026198 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.042800 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.057863 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.081973 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.095345 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.113662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.113736 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.113751 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.113779 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.113798 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.117637 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.216495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.216556 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.216569 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.216590 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.216605 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.244783 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.244922 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.244783 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:07 crc kubenswrapper[4888]: E1124 00:26:07.244977 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:07 crc kubenswrapper[4888]: E1124 00:26:07.245148 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:07 crc kubenswrapper[4888]: E1124 00:26:07.245327 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.319369 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.319424 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.319436 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.319456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.319468 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.422643 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.422702 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.422720 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.422748 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.422767 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.526928 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.526993 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.527009 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.527030 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.527042 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.630174 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.630250 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.630270 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.630297 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.630316 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.733220 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.733305 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.733337 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.733376 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.733397 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.790506 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/0.log" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.790587 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerStarted","Data":"16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.815744 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.831198 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.836223 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.836253 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.836265 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.836286 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.836300 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.849361 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.863662 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.882547 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.900548 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.920685 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.939156 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.939199 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.939210 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.939229 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.939240 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:07Z","lastTransitionTime":"2025-11-24T00:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.940134 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.957373 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.970777 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:07 crc kubenswrapper[4888]: I1124 00:26:07.985387 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:07Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.012859 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.031703 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.042220 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.042285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.042304 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.042331 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.042351 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.050583 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.064159 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.079892 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.095039 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.110548 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.121879 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.144349 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.144380 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.144392 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.144408 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.144419 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.244535 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:08 crc kubenswrapper[4888]: E1124 00:26:08.244832 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.249610 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.249678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.249692 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.249714 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.249730 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.268420 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.282774 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.299508 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.315552 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.333670 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.347305 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.351568 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.351616 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.351628 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.351644 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.351657 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.373510 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.392293 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.414592 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.432620 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.444891 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.454807 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.454925 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.454948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.454978 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.454997 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.459268 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.475611 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.492033 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.504063 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.523549 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.543212 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.558189 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.558227 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.558241 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.558263 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.558279 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.565794 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.580295 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:08Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.660786 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.660865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.660882 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.660908 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.660924 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.764017 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.764095 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.764115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.764144 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.764166 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.867151 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.867233 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.867253 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.867285 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.867309 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.970905 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.970967 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.970983 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.971007 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:08 crc kubenswrapper[4888]: I1124 00:26:08.971027 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:08Z","lastTransitionTime":"2025-11-24T00:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.074386 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.074443 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.074458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.074480 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.074497 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.177609 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.177658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.177667 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.177691 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.177701 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.244984 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.245049 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:09 crc kubenswrapper[4888]: E1124 00:26:09.245171 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.245239 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:09 crc kubenswrapper[4888]: E1124 00:26:09.245249 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:09 crc kubenswrapper[4888]: E1124 00:26:09.245475 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.280255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.280431 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.280452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.280474 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.280503 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.384294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.384378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.384406 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.384446 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.384474 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.488085 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.488139 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.488153 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.488175 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.488195 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.591353 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.591428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.591452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.591481 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.591502 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.694831 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.694899 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.694917 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.694943 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.694962 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.808948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.809003 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.809016 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.809042 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.809059 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.912583 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.912632 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.912645 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.912662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:09 crc kubenswrapper[4888]: I1124 00:26:09.912675 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:09Z","lastTransitionTime":"2025-11-24T00:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.015769 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.015850 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.015865 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.015885 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.015900 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.119473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.119526 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.119536 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.119553 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.119563 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.222629 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.222689 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.222703 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.222723 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.222744 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.244286 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:10 crc kubenswrapper[4888]: E1124 00:26:10.244511 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.325893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.325960 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.325977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.326006 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.326029 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.429013 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.429091 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.429103 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.429155 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.429169 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.532327 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.532394 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.532405 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.532427 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.532439 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.635720 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.635789 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.635802 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.635854 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.635874 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.738876 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.738936 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.738945 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.738963 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.738974 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.841693 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.841760 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.841780 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.841844 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.841874 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.944947 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.945022 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.945036 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.945055 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:10 crc kubenswrapper[4888]: I1124 00:26:10.945067 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:10Z","lastTransitionTime":"2025-11-24T00:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.048008 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.048056 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.048068 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.048086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.048098 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.151034 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.151104 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.151122 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.151148 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.151162 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.245246 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.245317 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.245421 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:11 crc kubenswrapper[4888]: E1124 00:26:11.245528 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:11 crc kubenswrapper[4888]: E1124 00:26:11.245704 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:11 crc kubenswrapper[4888]: E1124 00:26:11.245892 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.253456 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.253495 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.253530 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.253549 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.253560 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.356580 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.356632 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.356646 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.356664 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.356678 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.459599 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.459658 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.459678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.459704 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.459722 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.563323 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.563381 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.563396 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.563419 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.563439 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.666984 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.667075 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.667113 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.667145 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.667171 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.769734 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.769784 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.769799 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.769841 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.769858 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.872435 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.872491 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.872502 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.872520 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.872532 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.975011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.975042 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.975054 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.975068 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:11 crc kubenswrapper[4888]: I1124 00:26:11.975079 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:11Z","lastTransitionTime":"2025-11-24T00:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.078367 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.078431 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.078452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.078478 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.078496 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.182432 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.182477 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.182490 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.182512 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.182527 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.244623 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:12 crc kubenswrapper[4888]: E1124 00:26:12.244887 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.286003 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.286047 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.286083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.286107 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.286126 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.388973 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.389040 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.389083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.389131 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.389146 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.491891 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.491944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.491959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.491976 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.491987 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.594616 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.594651 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.594662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.594681 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.594694 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.697675 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.697866 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.697878 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.697895 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.697908 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.800118 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.800162 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.800172 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.800188 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.800199 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.902744 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.902886 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.902909 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.902938 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:12 crc kubenswrapper[4888]: I1124 00:26:12.902959 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:12Z","lastTransitionTime":"2025-11-24T00:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.006861 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.006944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.006966 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.006995 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.007017 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.110940 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.111064 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.111087 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.111111 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.111127 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.214449 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.214498 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.214517 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.214539 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.214555 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.244224 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.244313 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:13 crc kubenswrapper[4888]: E1124 00:26:13.244365 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.244391 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:13 crc kubenswrapper[4888]: E1124 00:26:13.244519 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:13 crc kubenswrapper[4888]: E1124 00:26:13.244685 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.326190 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.326252 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.326267 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.326296 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.326313 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.429834 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.429879 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.429893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.429915 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.429929 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.533764 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.533803 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.533832 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.533850 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.533862 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.636826 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.636863 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.636872 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.636888 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.636897 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.740602 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.740657 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.740675 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.740704 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.740725 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.844276 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.844336 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.844353 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.844377 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.844395 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.948288 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.948366 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.948389 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.948418 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:13 crc kubenswrapper[4888]: I1124 00:26:13.948437 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:13Z","lastTransitionTime":"2025-11-24T00:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.052514 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.052587 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.052598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.052617 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.052635 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.155855 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.155925 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.155943 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.155971 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.155992 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.245379 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:14 crc kubenswrapper[4888]: E1124 00:26:14.245717 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.259652 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.259738 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.259770 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.259804 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.259864 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.362648 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.362706 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.362725 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.362752 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.362770 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.466365 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.466552 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.466581 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.466649 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.466667 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.570346 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.570404 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.570421 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.570442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.570457 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.673608 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.673685 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.673710 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.673741 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.673766 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.777136 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.777475 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.777539 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.777679 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.777759 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.881197 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.881267 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.881287 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.881325 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.881345 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.984432 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.984490 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.984502 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.984521 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:14 crc kubenswrapper[4888]: I1124 00:26:14.984535 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:14Z","lastTransitionTime":"2025-11-24T00:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.033961 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.034028 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.034042 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.034062 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.034078 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.050775 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:15Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.055872 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.055932 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.055961 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.055990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.056010 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.078228 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:15Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.084268 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.084333 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.084355 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.084386 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.084407 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.105389 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:15Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.111621 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.111688 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.111712 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.111742 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.111761 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.131241 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:15Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.137302 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.137377 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.137396 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.137428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.137449 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.166515 4888 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2aca63cc-91d1-4463-bed3-9c285b4b381e\\\",\\\"systemUUID\\\":\\\"99fa3412-cc6b-4ec6-b00c-fbecc545cd5b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:15Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.166771 4888 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.169981 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.170060 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.170091 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.170170 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.170199 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.244864 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.244987 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.245084 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.245225 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.245526 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:15 crc kubenswrapper[4888]: E1124 00:26:15.245598 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.274426 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.274499 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.274524 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.274557 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.274576 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.378093 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.378195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.378222 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.378255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.378286 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.480900 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.481007 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.481036 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.481067 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.481088 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.583886 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.583941 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.583956 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.583978 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.583994 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.687257 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.687324 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.687336 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.687358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.687372 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.790798 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.790907 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.790929 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.790957 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.790977 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.894432 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.894535 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.894560 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.894597 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.894622 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.998053 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.998101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.998111 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.998131 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:15 crc kubenswrapper[4888]: I1124 00:26:15.998145 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:15Z","lastTransitionTime":"2025-11-24T00:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.102015 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.102074 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.102090 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.102113 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.102130 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.205473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.205537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.205548 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.205566 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.205578 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.244716 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:16 crc kubenswrapper[4888]: E1124 00:26:16.244969 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.310192 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.310284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.310311 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.310358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.310381 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.413442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.413494 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.413512 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.413537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.413549 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.517365 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.517404 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.517416 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.517435 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.517446 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.620899 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.620966 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.620983 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.621011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.621027 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.724552 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.724643 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.724660 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.724682 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.724699 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.827216 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.827263 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.827274 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.827294 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.827306 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.930135 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.930182 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.930195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.930214 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:16 crc kubenswrapper[4888]: I1124 00:26:16.930229 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:16Z","lastTransitionTime":"2025-11-24T00:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.032947 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.033019 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.033030 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.033050 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.033061 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.136891 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.136977 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.136998 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.137029 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.137051 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.240648 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.240703 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.240715 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.240737 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.240752 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.245214 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.245261 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.245496 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:17 crc kubenswrapper[4888]: E1124 00:26:17.245486 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:17 crc kubenswrapper[4888]: E1124 00:26:17.245715 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:17 crc kubenswrapper[4888]: E1124 00:26:17.245895 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.247005 4888 scope.go:117] "RemoveContainer" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.343851 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.343908 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.343924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.343951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.343967 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.447881 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.447934 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.447944 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.447963 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.447972 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.550483 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.550537 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.550569 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.550598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.550614 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.656224 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.656276 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.656289 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.656311 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.656332 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.759166 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.759242 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.759264 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.759296 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.759318 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.841265 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/2.log" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.843249 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.844637 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.858415 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.861442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.861467 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.861476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.861490 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.861499 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.878788 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.893408 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.909056 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.923948 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.939610 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.953494 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.966055 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.966104 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.966115 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.966140 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.966152 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:17Z","lastTransitionTime":"2025-11-24T00:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.970693 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:17 crc kubenswrapper[4888]: I1124 00:26:17.985195 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:17Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.008303 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.020382 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.036183 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.047390 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.062913 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.068615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.068667 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.068678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.068698 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.068711 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.080750 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.092920 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.108261 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.121279 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.131747 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.171924 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.172388 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.172400 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.172415 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.172427 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.244333 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:18 crc kubenswrapper[4888]: E1124 00:26:18.244477 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.256737 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.268332 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.275451 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.275531 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.275555 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.275581 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.275600 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.285203 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.296023 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.305148 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.313925 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.323143 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.335115 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.357163 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.371362 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.380120 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.380185 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.380196 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.380217 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.380234 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.410351 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.436173 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.458920 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.473237 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.482908 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.482948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.482961 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.482981 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.482993 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.488274 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.502177 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.522321 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.536855 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.551036 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.585961 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.586014 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.586024 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.586040 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.586055 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.688862 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.688950 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.688975 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.689016 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.689038 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.791367 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.791428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.791444 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.791466 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.791487 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.852165 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/3.log" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.852924 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/2.log" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.856153 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" exitCode=1 Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.856216 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.856287 4888 scope.go:117] "RemoveContainer" containerID="dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.857461 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:26:18 crc kubenswrapper[4888]: E1124 00:26:18.857775 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.883486 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.895255 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.895321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.895346 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.895375 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.895395 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.906379 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.930188 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.946304 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.963039 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.979308 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.993321 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:18Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.997951 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.998000 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.998015 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.998037 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:18 crc kubenswrapper[4888]: I1124 00:26:18.998052 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:18Z","lastTransitionTime":"2025-11-24T00:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.006964 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.030118 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.054590 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.081417 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.101650 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.101947 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.102044 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.102135 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.102173 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.110356 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd6ce54328529f8e0ad033083ec0fc9f62824477d915cc1ef7453701f7a3fcc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:25:46Z\\\",\\\"message\\\":\\\"ups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 00:25:46.265895 6528 services_controller.go:452] Built service openshift-operator-lifecycle-manager/packageserver-service per-node LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265901 6528 services_controller.go:453] Built service openshift-operator-lifecycle-manager/packageserver-service template LB for network=default: []services.LB{}\\\\nI1124 00:25:46.265909 6528 services_controller.go:454] Service openshift-operator-lifecycle-manager/packageserver-service for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1124 00:25:46.265922 6528 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster\\\\\\\", UUID:\\\\\\\"5e50827b-d271-442b-b8a7-7f33b2cd6b11\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/packageserver-service\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:18Z\\\",\\\"message\\\":\\\":18.093356 6884 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 00:26:18.093542 6884 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 00:26:18.093619 6884 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 00:26:18.093635 6884 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 00:26:18.093708 6884 factory.go:656] Stopping watch factory\\\\nI1124 00:26:18.093798 6884 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 00:26:18.093878 6884 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 00:26:18.094158 6884 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 00:26:18.094718 6884 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:26:18.094833 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 00:26:18.095029 6884 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.130364 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.154046 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.168045 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.188419 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.205528 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.205593 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.205615 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.205646 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.205664 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.208367 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.231074 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.244709 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.244792 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:19 crc kubenswrapper[4888]: E1124 00:26:19.245022 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.245204 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:19 crc kubenswrapper[4888]: E1124 00:26:19.245310 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:19 crc kubenswrapper[4888]: E1124 00:26:19.245461 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.251640 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.308681 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.308733 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.308744 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.308763 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.308776 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.412169 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.412223 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.412236 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.412253 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.412265 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.515366 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.515414 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.515429 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.515448 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.515461 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.618444 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.618501 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.618514 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.618555 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.618568 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.721585 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.722064 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.722207 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.722392 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.722528 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.826262 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.826311 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.826340 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.826358 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.826369 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.863256 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/3.log" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.868472 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:26:19 crc kubenswrapper[4888]: E1124 00:26:19.869054 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.890705 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8c6d153-8965-48cc-ac5f-850501f18690\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbe30653420817fb026739d90aa7111d0ee6025404c7c92392055bdf37715c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcebc46823763d554c57efe4ce0fd3ed4e51bf64839e09851a757598365c8994\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d2c55194773b31bf99403da32adcd67cca9dc4f394dc9b929a748e92795dee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d70c1c5a9a747ec50c9cd80a611f4ca726571733baa5a3c3c995d34c80860\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8984458f62e23229e1049d5ce7d7c054c45cc42f0e2d9bc922f6d1be66e492cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d0bd65e32f2fe32268ad79daccc2255deaf497a4befebede46bf776bcab14cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.911021 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0313c3c8-730e-4980-9164-0c3a77bc87c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d895a7e98d135880f405defc364ec90843b48c62890d3902e6f519311e4f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ad8af736f69efd6e2dd464e0c98aa0abd8809111219f7460240c94167b08aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291898aea1ac3b84948855250166d898308a2200c62951b9e1b682d3b72eb3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fac78c75e17fb4213c2e9e3c266e7f0d05fce9ed9f3331822cbe96e9d1428e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.928964 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.929035 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.929054 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.929078 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.929095 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:19Z","lastTransitionTime":"2025-11-24T00:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.933244 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab8b403add1cb79a07ae0a7efc0fb8200eb482946ad5365f5bd409e4c417a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d92bb45d8d3a0702a731eefab2ab06813c4cb6994409bf8ca92cb4576ea745b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.952876 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.973740 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zbfxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ba4b666-4513-46b8-b0f8-f39f56fb558f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:06Z\\\",\\\"message\\\":\\\"2025-11-24T00:25:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8\\\\n2025-11-24T00:25:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d64b4538-301f-49f5-a878-2097c5b7afc8 to /host/opt/cni/bin/\\\\n2025-11-24T00:25:21Z [verbose] multus-daemon started\\\\n2025-11-24T00:25:21Z [verbose] Readiness Indicator file check\\\\n2025-11-24T00:26:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:26:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ghmx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zbfxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:19 crc kubenswrapper[4888]: I1124 00:26:19.993516 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee556f32-31ee-45ef-bf08-26bf19390c96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whnmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7j4n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:19Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.030168 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9563d8ad-35d8-4a93-a5c9-37a94cda7421\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b8064e23e6bc98099681ab6b2c083bf58aa2a8578364925421b32eb2c533e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b957e74969883cd01d64dc2a80c64b082d229909384bf1afc2c2784f7243f3d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed94bd19f74e6d25acc275fa5824c18d8073e4d6b188a0ad094fae8f958369eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df221268e940434e735286c704d25b53a22afe0547c83ee4e1f689f0909921d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3244fa1a14d267dfb85af9c225df958d089bf6725a6cbd90b7e6a5a3d1dff981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234afb9959fb42282998c8386271ab6bcb8103ac84495781808b2f2cf81392b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdefc0fe1cc6a64e6d7566f04c61b1d61f365d8a498d3f294fcb7757ae0f295d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436888bf4efda6bb4d927cc19d58489f4844873d48d61f833d4a50fcade332ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.037435 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.037484 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.037499 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.037520 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.037533 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.053614 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.072000 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ltl44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adcded18-1bca-43c9-ac95-40ba409b3549\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de082506ce3c7a28e48f2eea07e3cb792a6ddf4965d1efda5cd6d4a1bf84242\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70128517eb9774dfafd3aef9d3b787159a9c0d5f955b33523d3887f8dfece732\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38faa56b2efe8e565152e7cec256f6c90e3a16941aaeed0a1e9c894d0bf3a67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c171201d12a25d95ca504443f61f234faa95f2085ccebeef7ff471a4914db38a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ed94cf1a1334068ba0779d2eb7d341e779bd5e27f4f75ae180a01881321ab09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faac3bf9feb7dbff73fd87d2fe92e8848f384afaa1cd7949950d87e559b59e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a903da69fff0e471d2b65365a10e8e56767c24d2bce93cacfb43163a18ab3cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xns87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ltl44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.090074 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ff1601-c0a4-405b-bbc3-7294295557e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c9408edcb33a2c4c7fd990efa22ab2a11afa0eb9875d836007789a621991b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4839a97f8deea70e5308d215445a9faea318037c11d5ef8d4e1b2a5599c5ef00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.113980 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34feb2ce-6980-4573-a805-541aa365640c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847ee60b3a6813c9065437f4ce52e25899312fa4a912dcb0dc05b2177b96fd33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac996d8f4c464790090693e18da10460ce7e4c34bb8b1a3d098585669c9edb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0abf46f0d93fd021f2fa86bb3603fbaf30713b957749bfdf5bf68f07a3835859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08eabc6dce3ce8c2e3c9ec283233e399ce99b86fa434e157f61af4a8de688e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:24:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.129884 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.140606 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.140671 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.140695 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.140722 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.140746 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.145224 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ead1da89da7c178f256041f3e1a47e5794607a462998227bb9c1d11c4d7898a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.161695 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dpn9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a234307-6086-494b-9373-19b006699223\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777d3f86f2920eefcd53ce715abf84a88a1c8db5a8e207c31d6ee6509c28bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r978x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dpn9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.176184 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hwxtq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8109bc-1f84-4f79-96e3-a74cf1b1041c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6079d456a31c00c45ac8c42e5a692f112dfa507eddf0982ab18769c779a82bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tl7gv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hwxtq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.189680 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d17c94b3-864e-4594-8ed7-06819e60fa5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e5d6dee8fe148cb6ea8d90245586a012b29e44f3f2cbd8988da4b0e4f101fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tkt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wvkcz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.206907 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2072a40fa3e39d1b04470233b6a163bf6106f18bbc3679d34462a831be9abfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.227616 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T00:26:18Z\\\",\\\"message\\\":\\\":18.093356 6884 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 00:26:18.093542 6884 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 00:26:18.093619 6884 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 00:26:18.093635 6884 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 00:26:18.093708 6884 factory.go:656] Stopping watch factory\\\\nI1124 00:26:18.093798 6884 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 00:26:18.093878 6884 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 00:26:18.094158 6884 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 00:26:18.094718 6884 ovnkube.go:599] Stopped ovnkube\\\\nI1124 00:26:18.094833 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 00:26:18.095029 6884 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T00:26:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T00:25:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T00:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mz76r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6j5lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.241528 4888 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"679c60ec-6115-419c-9074-da8a6e01b748\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T00:25:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eae6e29ec3907f08188ddba48b33b74ff175212a8a3865e989ecb263ed6f223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16b97cd7f71991cf0438359f3c0442f28c163e2e2ce13ddc5d96a175948a04f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T00:25:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plrm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T00:25:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k6trs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T00:26:20Z is after 2025-08-24T17:21:41Z" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.243228 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.243284 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.243296 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.243321 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.243333 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.244172 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:20 crc kubenswrapper[4888]: E1124 00:26:20.244296 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.346912 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.346971 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.346985 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.347004 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.347015 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.450256 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.450315 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.450327 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.450350 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.450371 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.553986 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.554083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.554096 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.554114 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.554150 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.657933 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.657990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.658001 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.658019 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.658033 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.760923 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.760990 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.761011 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.761038 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.761056 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.865018 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.865083 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.865096 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.865117 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.865132 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.968713 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.968799 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.968853 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.968881 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:20 crc kubenswrapper[4888]: I1124 00:26:20.968899 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:20Z","lastTransitionTime":"2025-11-24T00:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.017782 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.017968 4888 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.018064 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.018036561 +0000 UTC m=+147.600720625 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.071868 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.071933 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.071948 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.071973 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.071993 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.118895 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.119031 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.119076 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.119105 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119187 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.119145676 +0000 UTC m=+147.701829730 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119270 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119283 4888 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119467 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.119430724 +0000 UTC m=+147.702114798 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119296 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119548 4888 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119309 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119612 4888 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119624 4888 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119640 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.119610579 +0000 UTC m=+147.702294663 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.119666 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.11965436 +0000 UTC m=+147.702338434 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.175354 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.175442 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.175461 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.175493 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.175512 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.244753 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.244794 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.244972 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.245218 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.245325 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:21 crc kubenswrapper[4888]: E1124 00:26:21.245516 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.282338 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.282434 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.282455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.282483 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.282506 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.385542 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.385613 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.385634 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.385673 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.385698 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.489659 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.489718 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.489737 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.489771 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.489796 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.593428 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.593526 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.593546 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.593600 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.593616 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.697762 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.697899 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.697921 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.697952 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.697975 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.802562 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.802639 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.802662 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.802695 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.802719 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.905698 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.905746 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.905758 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.905781 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:21 crc kubenswrapper[4888]: I1124 00:26:21.905794 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:21Z","lastTransitionTime":"2025-11-24T00:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.009517 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.009570 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.009593 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.009621 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.009640 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.113308 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.113656 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.113666 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.113688 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.113701 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.218027 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.218086 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.218099 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.218121 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.218140 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.245092 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:22 crc kubenswrapper[4888]: E1124 00:26:22.245388 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.321386 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.321444 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.321461 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.321489 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.321507 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.424007 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.424069 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.424084 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.424104 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.424118 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.527719 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.527793 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.527830 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.527851 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.527865 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.631231 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.631275 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.631286 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.631304 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.631344 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.734095 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.734151 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.734167 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.734194 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.734211 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.837382 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.837455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.837476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.837504 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.837527 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.941103 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.941155 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.941172 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.941195 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:22 crc kubenswrapper[4888]: I1124 00:26:22.941214 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:22Z","lastTransitionTime":"2025-11-24T00:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.044501 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.044557 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.044569 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.044589 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.044602 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.147242 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.147291 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.147302 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.147323 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.147335 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.244648 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.244649 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.244704 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:23 crc kubenswrapper[4888]: E1124 00:26:23.244897 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:23 crc kubenswrapper[4888]: E1124 00:26:23.244956 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:23 crc kubenswrapper[4888]: E1124 00:26:23.245051 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.249842 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.249879 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.249891 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.249909 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.249922 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.352929 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.352979 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.352988 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.353003 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.353013 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.455310 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.455384 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.455406 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.456982 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.457044 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.560378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.560445 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.560462 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.560488 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.560506 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.663898 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.663962 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.663975 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.663996 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.664009 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.768283 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.768357 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.768376 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.768405 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.768426 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.871380 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.871446 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.871464 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.871488 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.871505 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.975065 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.975133 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.975152 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.975183 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:23 crc kubenswrapper[4888]: I1124 00:26:23.975206 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:23Z","lastTransitionTime":"2025-11-24T00:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.078447 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.078502 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.078518 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.078539 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.078552 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.181386 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.181439 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.181458 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.181481 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.181494 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.244571 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:24 crc kubenswrapper[4888]: E1124 00:26:24.245211 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.284800 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.284896 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.284918 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.284942 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.284957 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.388247 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.388313 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.388328 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.388354 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.388376 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.492226 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.492372 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.492399 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.492447 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.492475 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.594893 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.594960 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.594972 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.594996 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.595008 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.698263 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.698338 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.698378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.698422 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.698446 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.801364 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.801455 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.801476 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.801500 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.801543 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.904959 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.905048 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.905067 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.905101 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:24 crc kubenswrapper[4888]: I1124 00:26:24.905124 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:24Z","lastTransitionTime":"2025-11-24T00:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.009483 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.009552 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.009573 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.009604 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.009624 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.112570 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.112644 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.112666 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.112694 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.112713 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.216410 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.216545 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.216567 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.216598 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.216621 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.244649 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.244703 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:25 crc kubenswrapper[4888]: E1124 00:26:25.244947 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.244966 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:25 crc kubenswrapper[4888]: E1124 00:26:25.245194 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:25 crc kubenswrapper[4888]: E1124 00:26:25.245358 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.320334 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.320429 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.320449 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.320480 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.320500 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.424385 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.424576 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.424618 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.424654 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.424681 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.527678 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.527768 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.527793 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.527894 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.527921 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.566378 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.566452 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.566473 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.566552 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.566574 4888 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T00:26:25Z","lastTransitionTime":"2025-11-24T00:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.641900 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w"] Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.645213 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.647897 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.648002 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.648169 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.652121 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.676035 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16fc6b4f-6d84-452c-acac-8730ccfe5130-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.676204 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.676242 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.676299 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16fc6b4f-6d84-452c-acac-8730ccfe5130-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.676331 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16fc6b4f-6d84-452c-acac-8730ccfe5130-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.714899 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=63.714876172 podStartE2EDuration="1m3.714876172s" podCreationTimestamp="2025-11-24 00:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.691906133 +0000 UTC m=+88.274590257" watchObservedRunningTime="2025-11-24 00:26:25.714876172 +0000 UTC m=+88.297560216" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.760502 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ltl44" podStartSLOduration=67.7604823 podStartE2EDuration="1m7.7604823s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.741686689 +0000 UTC m=+88.324370733" watchObservedRunningTime="2025-11-24 00:26:25.7604823 +0000 UTC m=+88.343166344" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.771679 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podStartSLOduration=67.771649215 podStartE2EDuration="1m7.771649215s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.761291342 +0000 UTC m=+88.343975386" watchObservedRunningTime="2025-11-24 00:26:25.771649215 +0000 UTC m=+88.354333259" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.777585 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.777802 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.777854 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.778065 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16fc6b4f-6d84-452c-acac-8730ccfe5130-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.778102 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16fc6b4f-6d84-452c-acac-8730ccfe5130-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.778150 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16fc6b4f-6d84-452c-acac-8730ccfe5130-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.778079 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16fc6b4f-6d84-452c-acac-8730ccfe5130-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.780140 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16fc6b4f-6d84-452c-acac-8730ccfe5130-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.793842 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.79379726 podStartE2EDuration="1m7.79379726s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.793489812 +0000 UTC m=+88.376173856" watchObservedRunningTime="2025-11-24 00:26:25.79379726 +0000 UTC m=+88.376481304" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.794054 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.794047277 podStartE2EDuration="21.794047277s" podCreationTimestamp="2025-11-24 00:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.771950663 +0000 UTC m=+88.354634748" watchObservedRunningTime="2025-11-24 00:26:25.794047277 +0000 UTC m=+88.376731321" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.794349 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16fc6b4f-6d84-452c-acac-8730ccfe5130-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.804738 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16fc6b4f-6d84-452c-acac-8730ccfe5130-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xcn6w\" (UID: \"16fc6b4f-6d84-452c-acac-8730ccfe5130\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.843519 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dpn9j" podStartSLOduration=67.843494664 podStartE2EDuration="1m7.843494664s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.842903087 +0000 UTC m=+88.425587141" watchObservedRunningTime="2025-11-24 00:26:25.843494664 +0000 UTC m=+88.426178698" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.856152 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hwxtq" podStartSLOduration=67.85612607 podStartE2EDuration="1m7.85612607s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.855311337 +0000 UTC m=+88.437995381" watchObservedRunningTime="2025-11-24 00:26:25.85612607 +0000 UTC m=+88.438810124" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.910439 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k6trs" podStartSLOduration=67.91038279200001 podStartE2EDuration="1m7.910382792s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.909618211 +0000 UTC m=+88.492302255" watchObservedRunningTime="2025-11-24 00:26:25.910382792 +0000 UTC m=+88.493066846" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.954491 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.954468437 podStartE2EDuration="1m8.954468437s" podCreationTimestamp="2025-11-24 00:25:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.941540372 +0000 UTC m=+88.524224416" watchObservedRunningTime="2025-11-24 00:26:25.954468437 +0000 UTC m=+88.537152491" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.972893 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" Nov 24 00:26:25 crc kubenswrapper[4888]: I1124 00:26:25.976437 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=31.976410767 podStartE2EDuration="31.976410767s" podCreationTimestamp="2025-11-24 00:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:25.956289809 +0000 UTC m=+88.538973873" watchObservedRunningTime="2025-11-24 00:26:25.976410767 +0000 UTC m=+88.559094821" Nov 24 00:26:25 crc kubenswrapper[4888]: W1124 00:26:25.988543 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16fc6b4f_6d84_452c_acac_8730ccfe5130.slice/crio-8dc4db23b89ee39f2be2e1a3dc1fd6ad17350f3071333e71ba2f875e7c1fdb79 WatchSource:0}: Error finding container 8dc4db23b89ee39f2be2e1a3dc1fd6ad17350f3071333e71ba2f875e7c1fdb79: Status 404 returned error can't find the container with id 8dc4db23b89ee39f2be2e1a3dc1fd6ad17350f3071333e71ba2f875e7c1fdb79 Nov 24 00:26:26 crc kubenswrapper[4888]: I1124 00:26:26.023638 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zbfxr" podStartSLOduration=68.02361553 podStartE2EDuration="1m8.02361553s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:26.007745112 +0000 UTC m=+88.590429166" watchObservedRunningTime="2025-11-24 00:26:26.02361553 +0000 UTC m=+88.606299584" Nov 24 00:26:26 crc kubenswrapper[4888]: I1124 00:26:26.244430 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:26 crc kubenswrapper[4888]: E1124 00:26:26.245265 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:26 crc kubenswrapper[4888]: I1124 00:26:26.896957 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" event={"ID":"16fc6b4f-6d84-452c-acac-8730ccfe5130","Type":"ContainerStarted","Data":"2b4cde4c6130d21bf984722288b12f07eb2c09b59c306d81123052403998177c"} Nov 24 00:26:26 crc kubenswrapper[4888]: I1124 00:26:26.897043 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" event={"ID":"16fc6b4f-6d84-452c-acac-8730ccfe5130","Type":"ContainerStarted","Data":"8dc4db23b89ee39f2be2e1a3dc1fd6ad17350f3071333e71ba2f875e7c1fdb79"} Nov 24 00:26:27 crc kubenswrapper[4888]: I1124 00:26:27.245134 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:27 crc kubenswrapper[4888]: I1124 00:26:27.245202 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:27 crc kubenswrapper[4888]: I1124 00:26:27.245273 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:27 crc kubenswrapper[4888]: E1124 00:26:27.245309 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:27 crc kubenswrapper[4888]: E1124 00:26:27.245447 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:27 crc kubenswrapper[4888]: E1124 00:26:27.245513 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:28 crc kubenswrapper[4888]: I1124 00:26:28.244697 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:28 crc kubenswrapper[4888]: E1124 00:26:28.245010 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:29 crc kubenswrapper[4888]: I1124 00:26:29.244833 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:29 crc kubenswrapper[4888]: I1124 00:26:29.244984 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:29 crc kubenswrapper[4888]: I1124 00:26:29.245206 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:29 crc kubenswrapper[4888]: E1124 00:26:29.245183 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:29 crc kubenswrapper[4888]: E1124 00:26:29.245270 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:29 crc kubenswrapper[4888]: E1124 00:26:29.245380 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:30 crc kubenswrapper[4888]: I1124 00:26:30.244455 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:30 crc kubenswrapper[4888]: E1124 00:26:30.245072 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:31 crc kubenswrapper[4888]: I1124 00:26:31.244195 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:31 crc kubenswrapper[4888]: E1124 00:26:31.244429 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:31 crc kubenswrapper[4888]: I1124 00:26:31.244874 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:31 crc kubenswrapper[4888]: I1124 00:26:31.244943 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:31 crc kubenswrapper[4888]: E1124 00:26:31.245547 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:31 crc kubenswrapper[4888]: E1124 00:26:31.245880 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:31 crc kubenswrapper[4888]: I1124 00:26:31.245991 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:26:31 crc kubenswrapper[4888]: E1124 00:26:31.246255 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:26:32 crc kubenswrapper[4888]: I1124 00:26:32.244986 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:32 crc kubenswrapper[4888]: E1124 00:26:32.245844 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:33 crc kubenswrapper[4888]: I1124 00:26:33.245043 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:33 crc kubenswrapper[4888]: I1124 00:26:33.245084 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:33 crc kubenswrapper[4888]: E1124 00:26:33.245289 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:33 crc kubenswrapper[4888]: I1124 00:26:33.245190 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:33 crc kubenswrapper[4888]: E1124 00:26:33.245372 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:33 crc kubenswrapper[4888]: E1124 00:26:33.245558 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:34 crc kubenswrapper[4888]: I1124 00:26:34.245127 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:34 crc kubenswrapper[4888]: E1124 00:26:34.245466 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:35 crc kubenswrapper[4888]: I1124 00:26:35.245108 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:35 crc kubenswrapper[4888]: I1124 00:26:35.245178 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:35 crc kubenswrapper[4888]: I1124 00:26:35.245178 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:35 crc kubenswrapper[4888]: E1124 00:26:35.245338 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:35 crc kubenswrapper[4888]: E1124 00:26:35.245451 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:35 crc kubenswrapper[4888]: E1124 00:26:35.245535 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:36 crc kubenswrapper[4888]: I1124 00:26:36.244874 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:36 crc kubenswrapper[4888]: E1124 00:26:36.245075 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:36 crc kubenswrapper[4888]: I1124 00:26:36.807885 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:36 crc kubenswrapper[4888]: E1124 00:26:36.808101 4888 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:26:36 crc kubenswrapper[4888]: E1124 00:26:36.808202 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs podName:ee556f32-31ee-45ef-bf08-26bf19390c96 nodeName:}" failed. No retries permitted until 2025-11-24 00:27:40.808170187 +0000 UTC m=+163.390854261 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs") pod "network-metrics-daemon-7j4n7" (UID: "ee556f32-31ee-45ef-bf08-26bf19390c96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 00:26:37 crc kubenswrapper[4888]: I1124 00:26:37.244370 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:37 crc kubenswrapper[4888]: I1124 00:26:37.244597 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:37 crc kubenswrapper[4888]: I1124 00:26:37.244604 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:37 crc kubenswrapper[4888]: E1124 00:26:37.245392 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:37 crc kubenswrapper[4888]: E1124 00:26:37.245579 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:37 crc kubenswrapper[4888]: E1124 00:26:37.245852 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:38 crc kubenswrapper[4888]: I1124 00:26:38.244298 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:38 crc kubenswrapper[4888]: E1124 00:26:38.247279 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:39 crc kubenswrapper[4888]: I1124 00:26:39.244164 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:39 crc kubenswrapper[4888]: I1124 00:26:39.244164 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:39 crc kubenswrapper[4888]: I1124 00:26:39.244387 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:39 crc kubenswrapper[4888]: E1124 00:26:39.244555 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:39 crc kubenswrapper[4888]: E1124 00:26:39.244752 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:39 crc kubenswrapper[4888]: E1124 00:26:39.245051 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:40 crc kubenswrapper[4888]: I1124 00:26:40.245237 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:40 crc kubenswrapper[4888]: E1124 00:26:40.245532 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:41 crc kubenswrapper[4888]: I1124 00:26:41.244823 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:41 crc kubenswrapper[4888]: E1124 00:26:41.244989 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:41 crc kubenswrapper[4888]: I1124 00:26:41.245164 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:41 crc kubenswrapper[4888]: I1124 00:26:41.245246 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:41 crc kubenswrapper[4888]: E1124 00:26:41.245408 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:41 crc kubenswrapper[4888]: E1124 00:26:41.245803 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:42 crc kubenswrapper[4888]: I1124 00:26:42.244649 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:42 crc kubenswrapper[4888]: E1124 00:26:42.244920 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:43 crc kubenswrapper[4888]: I1124 00:26:43.244447 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:43 crc kubenswrapper[4888]: I1124 00:26:43.244477 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:43 crc kubenswrapper[4888]: E1124 00:26:43.245379 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:43 crc kubenswrapper[4888]: E1124 00:26:43.245474 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:43 crc kubenswrapper[4888]: I1124 00:26:43.244573 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:43 crc kubenswrapper[4888]: E1124 00:26:43.245785 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:44 crc kubenswrapper[4888]: I1124 00:26:44.244983 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:44 crc kubenswrapper[4888]: E1124 00:26:44.245506 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:45 crc kubenswrapper[4888]: I1124 00:26:45.244594 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:45 crc kubenswrapper[4888]: I1124 00:26:45.244601 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:45 crc kubenswrapper[4888]: I1124 00:26:45.244667 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:45 crc kubenswrapper[4888]: E1124 00:26:45.245173 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:45 crc kubenswrapper[4888]: E1124 00:26:45.245280 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:45 crc kubenswrapper[4888]: I1124 00:26:45.245510 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:26:45 crc kubenswrapper[4888]: E1124 00:26:45.245874 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6j5lq_openshift-ovn-kubernetes(2d662e59-55e9-45fa-870d-ad8c51a7a0ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" Nov 24 00:26:45 crc kubenswrapper[4888]: E1124 00:26:45.246117 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:46 crc kubenswrapper[4888]: I1124 00:26:46.244206 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:46 crc kubenswrapper[4888]: E1124 00:26:46.244556 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:47 crc kubenswrapper[4888]: I1124 00:26:47.244962 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:47 crc kubenswrapper[4888]: I1124 00:26:47.245073 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:47 crc kubenswrapper[4888]: I1124 00:26:47.245126 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:47 crc kubenswrapper[4888]: E1124 00:26:47.245247 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:47 crc kubenswrapper[4888]: E1124 00:26:47.245163 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:47 crc kubenswrapper[4888]: E1124 00:26:47.245508 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:48 crc kubenswrapper[4888]: I1124 00:26:48.244765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:48 crc kubenswrapper[4888]: E1124 00:26:48.244922 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:49 crc kubenswrapper[4888]: I1124 00:26:49.244837 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:49 crc kubenswrapper[4888]: I1124 00:26:49.244870 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:49 crc kubenswrapper[4888]: I1124 00:26:49.244961 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:49 crc kubenswrapper[4888]: E1124 00:26:49.245103 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:49 crc kubenswrapper[4888]: E1124 00:26:49.245198 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:49 crc kubenswrapper[4888]: E1124 00:26:49.245452 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:50 crc kubenswrapper[4888]: I1124 00:26:50.245031 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:50 crc kubenswrapper[4888]: E1124 00:26:50.245345 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:51 crc kubenswrapper[4888]: I1124 00:26:51.244238 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:51 crc kubenswrapper[4888]: I1124 00:26:51.244439 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:51 crc kubenswrapper[4888]: E1124 00:26:51.244786 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:51 crc kubenswrapper[4888]: E1124 00:26:51.244976 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:51 crc kubenswrapper[4888]: I1124 00:26:51.245151 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:51 crc kubenswrapper[4888]: E1124 00:26:51.245247 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.245411 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:52 crc kubenswrapper[4888]: E1124 00:26:52.245665 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.995953 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/1.log" Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.996700 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/0.log" Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.996773 4888 generic.go:334] "Generic (PLEG): container finished" podID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" containerID="16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7" exitCode=1 Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.996872 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerDied","Data":"16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7"} Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.996990 4888 scope.go:117] "RemoveContainer" containerID="2893d49484a3f5d5f8b946512e23687c490383ba1caaceb2f4420d322c65a3c6" Nov 24 00:26:52 crc kubenswrapper[4888]: I1124 00:26:52.997627 4888 scope.go:117] "RemoveContainer" containerID="16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7" Nov 24 00:26:52 crc kubenswrapper[4888]: E1124 00:26:52.998014 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-zbfxr_openshift-multus(4ba4b666-4513-46b8-b0f8-f39f56fb558f)\"" pod="openshift-multus/multus-zbfxr" podUID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" Nov 24 00:26:53 crc kubenswrapper[4888]: I1124 00:26:53.032638 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xcn6w" podStartSLOduration=95.032607527 podStartE2EDuration="1m35.032607527s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:26.919169237 +0000 UTC m=+89.501853281" watchObservedRunningTime="2025-11-24 00:26:53.032607527 +0000 UTC m=+115.615291611" Nov 24 00:26:53 crc kubenswrapper[4888]: I1124 00:26:53.245001 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:53 crc kubenswrapper[4888]: I1124 00:26:53.245043 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:53 crc kubenswrapper[4888]: I1124 00:26:53.245108 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:53 crc kubenswrapper[4888]: E1124 00:26:53.245202 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:53 crc kubenswrapper[4888]: E1124 00:26:53.245361 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:53 crc kubenswrapper[4888]: E1124 00:26:53.245483 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:54 crc kubenswrapper[4888]: I1124 00:26:54.003932 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/1.log" Nov 24 00:26:54 crc kubenswrapper[4888]: I1124 00:26:54.245005 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:54 crc kubenswrapper[4888]: E1124 00:26:54.245553 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:55 crc kubenswrapper[4888]: I1124 00:26:55.244920 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:55 crc kubenswrapper[4888]: E1124 00:26:55.245145 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:55 crc kubenswrapper[4888]: I1124 00:26:55.245322 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:55 crc kubenswrapper[4888]: E1124 00:26:55.245450 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:55 crc kubenswrapper[4888]: I1124 00:26:55.245654 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:55 crc kubenswrapper[4888]: E1124 00:26:55.245908 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:56 crc kubenswrapper[4888]: I1124 00:26:56.245037 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:56 crc kubenswrapper[4888]: E1124 00:26:56.245197 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:57 crc kubenswrapper[4888]: I1124 00:26:57.245126 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:57 crc kubenswrapper[4888]: I1124 00:26:57.245177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:57 crc kubenswrapper[4888]: I1124 00:26:57.245386 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:57 crc kubenswrapper[4888]: E1124 00:26:57.245510 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:26:57 crc kubenswrapper[4888]: E1124 00:26:57.245669 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:57 crc kubenswrapper[4888]: E1124 00:26:57.245757 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:58 crc kubenswrapper[4888]: I1124 00:26:58.244209 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:26:58 crc kubenswrapper[4888]: E1124 00:26:58.245918 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:26:58 crc kubenswrapper[4888]: I1124 00:26:58.246933 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:26:58 crc kubenswrapper[4888]: E1124 00:26:58.276231 4888 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 00:26:58 crc kubenswrapper[4888]: E1124 00:26:58.359010 4888 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.020552 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/3.log" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.023703 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerStarted","Data":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.024186 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.057042 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podStartSLOduration=101.057016496 podStartE2EDuration="1m41.057016496s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:26:59.055771461 +0000 UTC m=+121.638455525" watchObservedRunningTime="2025-11-24 00:26:59.057016496 +0000 UTC m=+121.639700550" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.100217 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7j4n7"] Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.100662 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:26:59 crc kubenswrapper[4888]: E1124 00:26:59.100779 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.245175 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:26:59 crc kubenswrapper[4888]: E1124 00:26:59.245338 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:26:59 crc kubenswrapper[4888]: I1124 00:26:59.245376 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:26:59 crc kubenswrapper[4888]: E1124 00:26:59.245448 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:27:00 crc kubenswrapper[4888]: I1124 00:27:00.244669 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:00 crc kubenswrapper[4888]: I1124 00:27:00.244718 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:00 crc kubenswrapper[4888]: E1124 00:27:00.244857 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:27:00 crc kubenswrapper[4888]: E1124 00:27:00.244908 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:27:01 crc kubenswrapper[4888]: I1124 00:27:01.244342 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:01 crc kubenswrapper[4888]: I1124 00:27:01.244412 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:01 crc kubenswrapper[4888]: E1124 00:27:01.244516 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:27:01 crc kubenswrapper[4888]: E1124 00:27:01.244590 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:27:02 crc kubenswrapper[4888]: I1124 00:27:02.244603 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:02 crc kubenswrapper[4888]: I1124 00:27:02.244642 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:02 crc kubenswrapper[4888]: E1124 00:27:02.244930 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:27:02 crc kubenswrapper[4888]: E1124 00:27:02.245192 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:27:03 crc kubenswrapper[4888]: I1124 00:27:03.244283 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:03 crc kubenswrapper[4888]: I1124 00:27:03.244343 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:03 crc kubenswrapper[4888]: E1124 00:27:03.244429 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:27:03 crc kubenswrapper[4888]: E1124 00:27:03.244542 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:27:03 crc kubenswrapper[4888]: E1124 00:27:03.360575 4888 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:27:04 crc kubenswrapper[4888]: I1124 00:27:04.244136 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:04 crc kubenswrapper[4888]: I1124 00:27:04.244184 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:04 crc kubenswrapper[4888]: E1124 00:27:04.244368 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:27:04 crc kubenswrapper[4888]: E1124 00:27:04.244545 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:27:05 crc kubenswrapper[4888]: I1124 00:27:05.245210 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:05 crc kubenswrapper[4888]: I1124 00:27:05.245243 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:05 crc kubenswrapper[4888]: E1124 00:27:05.245451 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:27:05 crc kubenswrapper[4888]: E1124 00:27:05.245633 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:27:06 crc kubenswrapper[4888]: I1124 00:27:06.245072 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:06 crc kubenswrapper[4888]: E1124 00:27:06.245283 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:27:06 crc kubenswrapper[4888]: I1124 00:27:06.245589 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:06 crc kubenswrapper[4888]: E1124 00:27:06.246187 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:27:06 crc kubenswrapper[4888]: I1124 00:27:06.246198 4888 scope.go:117] "RemoveContainer" containerID="16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7" Nov 24 00:27:07 crc kubenswrapper[4888]: I1124 00:27:07.051033 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/1.log" Nov 24 00:27:07 crc kubenswrapper[4888]: I1124 00:27:07.051104 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerStarted","Data":"2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799"} Nov 24 00:27:07 crc kubenswrapper[4888]: I1124 00:27:07.244177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:07 crc kubenswrapper[4888]: I1124 00:27:07.244210 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:07 crc kubenswrapper[4888]: E1124 00:27:07.244349 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 00:27:07 crc kubenswrapper[4888]: E1124 00:27:07.244553 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 00:27:08 crc kubenswrapper[4888]: I1124 00:27:08.245198 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:08 crc kubenswrapper[4888]: I1124 00:27:08.245198 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:08 crc kubenswrapper[4888]: E1124 00:27:08.245537 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 00:27:08 crc kubenswrapper[4888]: E1124 00:27:08.245689 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7j4n7" podUID="ee556f32-31ee-45ef-bf08-26bf19390c96" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.244785 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.244785 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.247147 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.247390 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.248291 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 00:27:09 crc kubenswrapper[4888]: I1124 00:27:09.248408 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 00:27:10 crc kubenswrapper[4888]: I1124 00:27:10.244359 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:10 crc kubenswrapper[4888]: I1124 00:27:10.244771 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:10 crc kubenswrapper[4888]: I1124 00:27:10.246859 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 00:27:10 crc kubenswrapper[4888]: I1124 00:27:10.247407 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.804356 4888 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.839583 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29399040-nzntm"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.840115 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.843276 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.843470 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.843693 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q26kt"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.844742 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.844925 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.845971 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.846023 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.846691 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.851043 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.852950 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.853141 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.853280 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.853856 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856090 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856302 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856328 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856394 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856545 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.856564 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.857196 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.858721 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.858736 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.858918 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8s79"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.859574 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.861098 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.862236 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.862526 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.862783 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.862991 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.863564 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xz576"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.864140 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.865157 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.870358 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.870427 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.870535 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871705 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871730 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871779 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871786 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871805 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871808 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871859 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.871978 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.872336 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.872361 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.872690 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.873932 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.874392 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876183 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876301 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876633 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876658 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876676 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.876824 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.878387 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.878438 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.878472 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.879011 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.879320 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.879637 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.879937 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.880199 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.880342 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.880656 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.880893 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.881248 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.881395 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.881523 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.881684 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.881703 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sl4qh"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.882461 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.882979 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.884426 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.885000 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.888425 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.888697 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.893030 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.893273 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.893529 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.895603 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.896494 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.897996 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911097 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5f942"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911574 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911610 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911651 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911938 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.911988 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912003 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912071 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912082 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912200 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912262 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912351 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912355 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912439 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912549 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912769 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912856 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.912893 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.913196 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.913308 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.913846 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.914123 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.919754 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.923550 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.923781 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.924023 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.924356 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.924987 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.925350 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.925984 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.926857 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.926985 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.927681 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.928019 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.928434 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.928627 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.928850 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.929187 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.930162 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.934554 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.934683 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.935031 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.935261 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.935804 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.936345 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.936640 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.937206 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.938623 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.940458 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.940607 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmph"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.941354 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.941571 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.941791 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.942865 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.945707 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.946239 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.956257 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.956707 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.961641 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.963120 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.979029 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.979303 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.980745 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xfrjt"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.981163 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.981222 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.981181 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j422v"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.981790 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.982496 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29399040-nzntm"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.983536 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.984233 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.985677 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.988163 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q26kt"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.988220 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.988537 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.988918 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.992092 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.992857 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.993206 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.996548 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:27:16 crc kubenswrapper[4888]: I1124 00:27:16.997331 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001192 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-serving-cert\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001235 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-image-import-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001267 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mk98\" (UniqueName: \"kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001290 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-client\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001309 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001331 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-serving-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001366 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001388 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001410 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-audit\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001430 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001456 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001482 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcwnf\" (UniqueName: \"kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001513 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001540 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001566 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-encryption-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001606 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001630 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001653 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj8g8\" (UniqueName: \"kubernetes.io/projected/dec11a6c-0eb1-428a-b624-3558ffc42e1b-kube-api-access-nj8g8\") pod \"downloads-7954f5f757-5f942\" (UID: \"dec11a6c-0eb1-428a-b624-3558ffc42e1b\") " pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001676 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001698 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001726 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-audit-dir\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001756 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001779 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001803 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001873 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001898 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001922 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001943 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt9tr\" (UniqueName: \"kubernetes.io/projected/48e64e1d-4554-4633-a7be-970573d5ad06-kube-api-access-nt9tr\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001964 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.001988 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002008 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002039 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5qnx\" (UniqueName: \"kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002064 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-node-pullsecrets\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002085 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002107 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002130 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002151 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002171 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002192 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.002214 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.012019 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.012655 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88dzb"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.013760 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.014031 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.020443 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.021093 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-htnbg"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.021585 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.022013 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.022356 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.025061 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.026709 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c29ws"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.027428 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.029169 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.029735 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.030228 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.030221 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.032498 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.033013 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.038489 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.039901 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.040935 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.041061 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.044988 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.045241 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.046153 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xz576"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.046267 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.053919 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.054158 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8s79"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.058917 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-tpmqr"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.059754 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.064885 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.066118 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmph"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.092629 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.093912 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.095469 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.101057 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.102906 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103525 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103578 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103614 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-audit\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103647 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103676 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103698 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcwnf\" (UniqueName: \"kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103749 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-encryption-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103777 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103805 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103878 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103907 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj8g8\" (UniqueName: \"kubernetes.io/projected/dec11a6c-0eb1-428a-b624-3558ffc42e1b-kube-api-access-nj8g8\") pod \"downloads-7954f5f757-5f942\" (UID: \"dec11a6c-0eb1-428a-b624-3558ffc42e1b\") " pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103952 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.103983 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104009 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104038 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-audit-dir\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104069 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104105 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104134 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104166 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104195 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104224 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104250 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt9tr\" (UniqueName: \"kubernetes.io/projected/48e64e1d-4554-4633-a7be-970573d5ad06-kube-api-access-nt9tr\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104277 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104456 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104524 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5qnx\" (UniqueName: \"kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104560 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-node-pullsecrets\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104590 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104624 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104655 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104688 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104718 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104749 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104776 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104823 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-serving-cert\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104854 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-image-import-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104883 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mk98\" (UniqueName: \"kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104908 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-client\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.104987 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-serving-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.106358 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-audit\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.106496 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.108682 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-node-pullsecrets\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.108685 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sl4qh"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.109600 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.112969 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-serving-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.113291 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.122156 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.140896 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.141606 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.141706 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48e64e1d-4554-4633-a7be-970573d5ad06-audit-dir\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.142345 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.142842 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.144209 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.144211 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.144984 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.145224 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48e64e1d-4554-4633-a7be-970573d5ad06-image-import-ca\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.146079 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.146514 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.146887 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.146925 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.147459 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.147525 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-encryption-config\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.147624 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.147619 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.147884 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.148264 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.148735 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.149036 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.149079 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.149346 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.149749 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.150081 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.151253 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.151635 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.153330 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.154151 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.154891 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.156389 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.156769 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.158222 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j422v"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.159545 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.161173 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.162772 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-l99d8"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.163510 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.163612 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.163719 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.164489 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-etcd-client\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.166198 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xfrjt"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.168898 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.168938 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.171474 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.172677 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e64e1d-4554-4633-a7be-970573d5ad06-serving-cert\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.174106 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.176734 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5f942"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.176961 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.180707 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.180762 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tpmqr"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.181295 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.185542 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.186134 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88dzb"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.187705 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.188544 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.191163 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.193377 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c29ws"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.194699 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.196066 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-b52cm"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.197113 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.197323 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b52cm"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.199011 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kstd5"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.200437 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kstd5"] Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.200626 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.202139 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.222415 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.243115 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.262746 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.282552 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.303095 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.323340 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.343071 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.362553 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.383535 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.402756 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.425178 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.443478 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.462303 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.483002 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.502917 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.522553 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.543045 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.562195 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.582867 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.603229 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.622633 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.643016 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.663369 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.682163 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.703535 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.722765 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.742961 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.762444 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.783724 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.814557 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.822668 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.844127 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.864616 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.882580 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.903385 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.925832 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.942937 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.962684 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 00:27:17 crc kubenswrapper[4888]: I1124 00:27:17.982634 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.002762 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.023730 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.040876 4888 request.go:700] Waited for 1.015290085s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.042563 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.062726 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.085946 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.102596 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.123775 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.143613 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.163376 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.183542 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.203423 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.223489 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.243072 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.262875 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.283180 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.301676 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.321969 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.342446 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.364318 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.382659 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.403341 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.422828 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.442993 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.461780 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.482395 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.502449 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.522254 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.542507 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.562670 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.589791 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.602624 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.624255 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.642767 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.663082 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.682320 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.702570 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.722131 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.742576 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.821745 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5qnx\" (UniqueName: \"kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx\") pod \"oauth-openshift-558db77b4-st95l\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823533 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-config\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823602 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823642 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823692 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a5f48a-3697-4b07-8922-7007d5a888e2-serving-cert\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823737 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823775 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7f7eff25-428d-4192-a4db-2ba3145969a3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823839 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e71b63a8-b89b-4e4f-83a2-4fad8694e177-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823952 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.823997 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-auth-proxy-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824132 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj287\" (UniqueName: \"kubernetes.io/projected/7f7eff25-428d-4192-a4db-2ba3145969a3-kube-api-access-lj287\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824171 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmmll\" (UniqueName: \"kubernetes.io/projected/103c3930-7610-4c38-af13-cb0f06da2629-kube-api-access-lmmll\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824281 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824324 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824410 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824459 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-service-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824518 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/103c3930-7610-4c38-af13-cb0f06da2629-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824556 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824596 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824622 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-serving-cert\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824655 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824684 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-config\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824714 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4c7c\" (UniqueName: \"kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824745 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824792 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94c7a44d-4411-4588-a723-69464372ce7d-serving-cert\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824855 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824902 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6cfb\" (UniqueName: \"kubernetes.io/projected/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-kube-api-access-r6cfb\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824947 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-client\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.824987 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhh6t\" (UniqueName: \"kubernetes.io/projected/11833290-48d3-4875-a9a6-92b4c9a2a0c2-kube-api-access-hhh6t\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825020 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsflc\" (UniqueName: \"kubernetes.io/projected/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-kube-api-access-bsflc\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825088 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825158 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-kube-api-access-h2rrx\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825198 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvlgx\" (UniqueName: \"kubernetes.io/projected/94c7a44d-4411-4588-a723-69464372ce7d-kube-api-access-fvlgx\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825231 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825265 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f7eff25-428d-4192-a4db-2ba3145969a3-serving-cert\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825298 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpwgv\" (UniqueName: \"kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825360 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rf5x\" (UniqueName: \"kubernetes.io/projected/34a5f48a-3697-4b07-8922-7007d5a888e2-kube-api-access-8rf5x\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825393 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg2h4\" (UniqueName: \"kubernetes.io/projected/e71b63a8-b89b-4e4f-83a2-4fad8694e177-kube-api-access-rg2h4\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825422 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-encryption-config\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825458 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-images\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825485 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825525 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-png9r\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825555 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825654 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-trusted-ca\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825683 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-dir\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825715 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825763 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-policies\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825799 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-config\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.825993 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71b63a8-b89b-4e4f-83a2-4fad8694e177-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.826020 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11833290-48d3-4875-a9a6-92b4c9a2a0c2-machine-approver-tls\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.826078 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: E1124 00:27:18.826719 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.326700507 +0000 UTC m=+141.909384891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.851070 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt9tr\" (UniqueName: \"kubernetes.io/projected/48e64e1d-4554-4633-a7be-970573d5ad06-kube-api-access-nt9tr\") pod \"apiserver-76f77b778f-q26kt\" (UID: \"48e64e1d-4554-4633-a7be-970573d5ad06\") " pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.865454 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcwnf\" (UniqueName: \"kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf\") pod \"console-f9d7485db-jdqr8\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.887629 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb02b82-c40c-4e4b-99dc-7cccd4bd092e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zj2zq\" (UID: \"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.898315 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.901698 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mk98\" (UniqueName: \"kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98\") pod \"route-controller-manager-6576b87f9c-mc99g\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.919871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj8g8\" (UniqueName: \"kubernetes.io/projected/dec11a6c-0eb1-428a-b624-3558ffc42e1b-kube-api-access-nj8g8\") pod \"downloads-7954f5f757-5f942\" (UID: \"dec11a6c-0eb1-428a-b624-3558ffc42e1b\") " pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.923848 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.926957 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:18 crc kubenswrapper[4888]: E1124 00:27:18.927168 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.427136516 +0000 UTC m=+142.009820590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.927658 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-service-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.927862 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c70994e1-07d8-4989-9483-4dae00ee88a9-config\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928032 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6d39688c-5aad-461d-aa95-d13f144526a7-metrics-tls\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928285 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928428 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-mountpoint-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928584 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/103c3930-7610-4c38-af13-cb0f06da2629-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928728 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928924 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929116 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ce82c4-8c24-4b10-9116-83975ed9155a-config\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.928783 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-service-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929561 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d8003a3-3839-40a4-a432-1addc1dc99ee-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929690 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6027adf5-7232-462c-9371-47187637c34c-proxy-tls\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929761 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-metrics-certs\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929793 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt7fm\" (UniqueName: \"kubernetes.io/projected/814a1313-1c15-436c-9a1d-132859fabdea-kube-api-access-nt7fm\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929842 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv8g2\" (UniqueName: \"kubernetes.io/projected/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-kube-api-access-qv8g2\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929871 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsstn\" (UniqueName: \"kubernetes.io/projected/c1f8e203-e571-46db-9414-710646e64525-kube-api-access-tsstn\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929909 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929938 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.929996 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvlgx\" (UniqueName: \"kubernetes.io/projected/94c7a44d-4411-4588-a723-69464372ce7d-kube-api-access-fvlgx\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930095 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930097 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930309 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-kube-api-access-h2rrx\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930605 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpwgv\" (UniqueName: \"kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930750 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fccb657c-29dd-4ccf-a16c-6867e2fd2145-service-ca-bundle\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.930972 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-csi-data-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931125 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-config\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931307 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg2h4\" (UniqueName: \"kubernetes.io/projected/e71b63a8-b89b-4e4f-83a2-4fad8694e177-kube-api-access-rg2h4\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931459 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-encryption-config\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931584 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931147 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.931848 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-images\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932038 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-png9r\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932180 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932315 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-stats-auth\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz7n5\" (UniqueName: \"kubernetes.io/projected/1d8111bb-3da5-4474-aee1-efa4173d83ba-kube-api-access-sz7n5\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932602 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537827c-6019-4b2d-a15b-8a4d742b1664-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-metrics-tls\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.932922 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933101 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933267 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-policies\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933484 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933683 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933871 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b931aa03-f254-4fa4-8f2f-04b735ae01f5-tmpfs\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934025 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r5nx\" (UniqueName: \"kubernetes.io/projected/f4af2053-39df-4362-aacb-0cd7bc018d9c-kube-api-access-4r5nx\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934349 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c70994e1-07d8-4989-9483-4dae00ee88a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934573 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjkmz\" (UniqueName: \"kubernetes.io/projected/2bd354c8-cd5f-43a6-880f-694516d453f7-kube-api-access-cjkmz\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934744 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5hfj\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-kube-api-access-g5hfj\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934752 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: E1124 00:27:18.935046 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.434979311 +0000 UTC m=+142.017663385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934025 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.933030 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934052 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-images\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934944 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk498\" (UniqueName: \"kubernetes.io/projected/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-kube-api-access-dk498\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.935645 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkdmj\" (UniqueName: \"kubernetes.io/projected/3d8003a3-3839-40a4-a432-1addc1dc99ee-kube-api-access-wkdmj\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.935797 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-socket-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.935904 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4tc4\" (UniqueName: \"kubernetes.io/projected/6d39688c-5aad-461d-aa95-d13f144526a7-kube-api-access-h4tc4\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.935956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5vj\" (UniqueName: \"kubernetes.io/projected/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-kube-api-access-7b5vj\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.935999 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6027adf5-7232-462c-9371-47187637c34c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.934763 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-policies\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936057 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936131 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hndpl\" (UniqueName: \"kubernetes.io/projected/b931aa03-f254-4fa4-8f2f-04b735ae01f5-kube-api-access-hndpl\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936178 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x42f\" (UniqueName: \"kubernetes.io/projected/e537827c-6019-4b2d-a15b-8a4d742b1664-kube-api-access-6x42f\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936206 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b82bk\" (UniqueName: \"kubernetes.io/projected/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-kube-api-access-b82bk\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936290 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a5f48a-3697-4b07-8922-7007d5a888e2-serving-cert\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936319 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-plugins-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936373 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936400 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7f7eff25-428d-4192-a4db-2ba3145969a3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936431 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936461 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936487 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlzqf\" (UniqueName: \"kubernetes.io/projected/8a357524-1674-4bc5-9f2f-e310387ca2be-kube-api-access-tlzqf\") pod \"migrator-59844c95c7-cldc2\" (UID: \"8a357524-1674-4bc5-9f2f-e310387ca2be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936536 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-service-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936581 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936612 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/814a1313-1c15-436c-9a1d-132859fabdea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936641 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936665 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dznjq\" (UniqueName: \"kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936694 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2ce82c4-8c24-4b10-9116-83975ed9155a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936738 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmmll\" (UniqueName: \"kubernetes.io/projected/103c3930-7610-4c38-af13-cb0f06da2629-kube-api-access-lmmll\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936765 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936792 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqm2g\" (UniqueName: \"kubernetes.io/projected/ee78c68a-a580-49cf-85d4-e172ac11fbab-kube-api-access-tqm2g\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936844 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936871 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936901 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4733c1c8-0664-4ccd-b53c-890be583a891-signing-key\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936961 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cft8b\" (UniqueName: \"kubernetes.io/projected/6027adf5-7232-462c-9371-47187637c34c-kube-api-access-cft8b\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936987 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937014 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c70994e1-07d8-4989-9483-4dae00ee88a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937042 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-config\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937064 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-serving-cert\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937111 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94c7a44d-4411-4588-a723-69464372ce7d-serving-cert\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937140 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4c7c\" (UniqueName: \"kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937180 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937212 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n5xw\" (UniqueName: \"kubernetes.io/projected/fccb657c-29dd-4ccf-a16c-6867e2fd2145-kube-api-access-4n5xw\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937242 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937290 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937323 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-node-bootstrap-token\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937351 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-client\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937384 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6cfb\" (UniqueName: \"kubernetes.io/projected/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-kube-api-access-r6cfb\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937412 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhh6t\" (UniqueName: \"kubernetes.io/projected/11833290-48d3-4875-a9a6-92b4c9a2a0c2-kube-api-access-hhh6t\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937437 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsflc\" (UniqueName: \"kubernetes.io/projected/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-kube-api-access-bsflc\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937460 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee78c68a-a580-49cf-85d4-e172ac11fbab-serving-cert\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937488 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937524 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f7eff25-428d-4192-a4db-2ba3145969a3-serving-cert\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937555 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rf5x\" (UniqueName: \"kubernetes.io/projected/34a5f48a-3697-4b07-8922-7007d5a888e2-kube-api-access-8rf5x\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937578 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537827c-6019-4b2d-a15b-8a4d742b1664-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937612 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937683 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcnf5\" (UniqueName: \"kubernetes.io/projected/81eba091-1a6d-4dff-9b38-33145da7c8bd-kube-api-access-rcnf5\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937724 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-trusted-ca\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937752 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-dir\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937775 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ns44\" (UniqueName: \"kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937800 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee78c68a-a580-49cf-85d4-e172ac11fbab-config\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937864 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-images\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937886 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937935 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-config\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937959 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71b63a8-b89b-4e4f-83a2-4fad8694e177-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.937987 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11833290-48d3-4875-a9a6-92b4c9a2a0c2-machine-approver-tls\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938244 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-srv-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938248 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-config\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938277 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ce82c4-8c24-4b10-9116-83975ed9155a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938358 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938392 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-srv-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938422 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xngd\" (UniqueName: \"kubernetes.io/projected/4733c1c8-0664-4ccd-b53c-890be583a891-kube-api-access-4xngd\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938455 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-config-volume\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938492 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-config\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938520 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d8003a3-3839-40a4-a432-1addc1dc99ee-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938572 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938598 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2bd354c8-cd5f-43a6-880f-694516d453f7-proxy-tls\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938626 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-cert\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938690 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-webhook-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938716 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36c453df-4dc2-47d5-b66f-688c55b01763-trusted-ca\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938764 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-serving-cert\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938791 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938846 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e71b63a8-b89b-4e4f-83a2-4fad8694e177-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938873 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d8111bb-3da5-4474-aee1-efa4173d83ba-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938906 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ff2q\" (UniqueName: \"kubernetes.io/projected/71e1aa4d-7bb1-4621-aada-6fbd776fe776-kube-api-access-4ff2q\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938945 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-auth-proxy-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.938970 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-default-certificate\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939020 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj287\" (UniqueName: \"kubernetes.io/projected/7f7eff25-428d-4192-a4db-2ba3145969a3-kube-api-access-lj287\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939045 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-registration-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939067 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/36c453df-4dc2-47d5-b66f-688c55b01763-metrics-tls\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939108 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-certs\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939133 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939164 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-client\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939184 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4733c1c8-0664-4ccd-b53c-890be583a891-signing-cabundle\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.939955 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103c3930-7610-4c38-af13-cb0f06da2629-config\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.936734 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.941068 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.941167 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a5f48a-3697-4b07-8922-7007d5a888e2-serving-cert\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.942065 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-serving-cert\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.942324 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/103c3930-7610-4c38-af13-cb0f06da2629-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.942402 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11833290-48d3-4875-a9a6-92b4c9a2a0c2-auth-proxy-config\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.944274 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-audit-dir\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.944405 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.944294 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.944765 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-encryption-config\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.945038 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-config\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.945110 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7f7eff25-428d-4192-a4db-2ba3145969a3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.945167 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94c7a44d-4411-4588-a723-69464372ce7d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.945625 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34a5f48a-3697-4b07-8922-7007d5a888e2-trusted-ca\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.945921 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71b63a8-b89b-4e4f-83a2-4fad8694e177-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.946095 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.947244 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.947263 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94c7a44d-4411-4588-a723-69464372ce7d-serving-cert\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.948032 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.948242 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-etcd-client\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.948583 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.949220 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e71b63a8-b89b-4e4f-83a2-4fad8694e177-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.949526 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.949657 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11833290-48d3-4875-a9a6-92b4c9a2a0c2-machine-approver-tls\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.950456 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f7eff25-428d-4192-a4db-2ba3145969a3-serving-cert\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.952611 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.963547 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.978607 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:18 crc kubenswrapper[4888]: I1124 00:27:18.983006 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.001632 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.004556 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.021668 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.023507 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.024233 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.034417 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.042672 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049547 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.049764 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.549730364 +0000 UTC m=+142.132414408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049854 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n5xw\" (UniqueName: \"kubernetes.io/projected/fccb657c-29dd-4ccf-a16c-6867e2fd2145-kube-api-access-4n5xw\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049888 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049911 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-node-bootstrap-token\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049949 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee78c68a-a580-49cf-85d4-e172ac11fbab-serving-cert\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.049968 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050001 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537827c-6019-4b2d-a15b-8a4d742b1664-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050020 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnf5\" (UniqueName: \"kubernetes.io/projected/81eba091-1a6d-4dff-9b38-33145da7c8bd-kube-api-access-rcnf5\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050045 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ns44\" (UniqueName: \"kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050061 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee78c68a-a580-49cf-85d4-e172ac11fbab-config\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050075 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-images\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050092 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050110 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-srv-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050128 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ce82c4-8c24-4b10-9116-83975ed9155a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050143 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050159 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-srv-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050176 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xngd\" (UniqueName: \"kubernetes.io/projected/4733c1c8-0664-4ccd-b53c-890be583a891-kube-api-access-4xngd\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050190 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-config-volume\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050208 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d8003a3-3839-40a4-a432-1addc1dc99ee-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050226 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2bd354c8-cd5f-43a6-880f-694516d453f7-proxy-tls\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050244 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-cert\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050262 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-webhook-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050279 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36c453df-4dc2-47d5-b66f-688c55b01763-trusted-ca\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050296 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-serving-cert\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050313 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050329 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d8111bb-3da5-4474-aee1-efa4173d83ba-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050347 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ff2q\" (UniqueName: \"kubernetes.io/projected/71e1aa4d-7bb1-4621-aada-6fbd776fe776-kube-api-access-4ff2q\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050371 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-default-certificate\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050394 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-registration-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050410 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/36c453df-4dc2-47d5-b66f-688c55b01763-metrics-tls\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050428 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-certs\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-client\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050464 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4733c1c8-0664-4ccd-b53c-890be583a891-signing-cabundle\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050482 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c70994e1-07d8-4989-9483-4dae00ee88a9-config\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050499 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6d39688c-5aad-461d-aa95-d13f144526a7-metrics-tls\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050516 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-mountpoint-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050532 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ce82c4-8c24-4b10-9116-83975ed9155a-config\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050572 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d8003a3-3839-40a4-a432-1addc1dc99ee-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050587 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6027adf5-7232-462c-9371-47187637c34c-proxy-tls\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050609 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt7fm\" (UniqueName: \"kubernetes.io/projected/814a1313-1c15-436c-9a1d-132859fabdea-kube-api-access-nt7fm\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050626 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv8g2\" (UniqueName: \"kubernetes.io/projected/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-kube-api-access-qv8g2\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050643 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-metrics-certs\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050663 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsstn\" (UniqueName: \"kubernetes.io/projected/c1f8e203-e571-46db-9414-710646e64525-kube-api-access-tsstn\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050682 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050720 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fccb657c-29dd-4ccf-a16c-6867e2fd2145-service-ca-bundle\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050743 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-csi-data-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050758 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-config\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050789 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-stats-auth\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050804 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz7n5\" (UniqueName: \"kubernetes.io/projected/1d8111bb-3da5-4474-aee1-efa4173d83ba-kube-api-access-sz7n5\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050821 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537827c-6019-4b2d-a15b-8a4d742b1664-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050840 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-metrics-tls\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050886 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050903 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b931aa03-f254-4fa4-8f2f-04b735ae01f5-tmpfs\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050920 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r5nx\" (UniqueName: \"kubernetes.io/projected/f4af2053-39df-4362-aacb-0cd7bc018d9c-kube-api-access-4r5nx\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050935 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c70994e1-07d8-4989-9483-4dae00ee88a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050953 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjkmz\" (UniqueName: \"kubernetes.io/projected/2bd354c8-cd5f-43a6-880f-694516d453f7-kube-api-access-cjkmz\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050976 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5hfj\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-kube-api-access-g5hfj\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.050992 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk498\" (UniqueName: \"kubernetes.io/projected/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-kube-api-access-dk498\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051007 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkdmj\" (UniqueName: \"kubernetes.io/projected/3d8003a3-3839-40a4-a432-1addc1dc99ee-kube-api-access-wkdmj\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051027 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4tc4\" (UniqueName: \"kubernetes.io/projected/6d39688c-5aad-461d-aa95-d13f144526a7-kube-api-access-h4tc4\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051067 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5vj\" (UniqueName: \"kubernetes.io/projected/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-kube-api-access-7b5vj\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051083 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6027adf5-7232-462c-9371-47187637c34c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051099 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-socket-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051118 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x42f\" (UniqueName: \"kubernetes.io/projected/e537827c-6019-4b2d-a15b-8a4d742b1664-kube-api-access-6x42f\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051134 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b82bk\" (UniqueName: \"kubernetes.io/projected/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-kube-api-access-b82bk\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051152 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hndpl\" (UniqueName: \"kubernetes.io/projected/b931aa03-f254-4fa4-8f2f-04b735ae01f5-kube-api-access-hndpl\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051182 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-plugins-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051208 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051225 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051242 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlzqf\" (UniqueName: \"kubernetes.io/projected/8a357524-1674-4bc5-9f2f-e310387ca2be-kube-api-access-tlzqf\") pod \"migrator-59844c95c7-cldc2\" (UID: \"8a357524-1674-4bc5-9f2f-e310387ca2be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051269 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-service-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051287 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/814a1313-1c15-436c-9a1d-132859fabdea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051306 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051322 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dznjq\" (UniqueName: \"kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2ce82c4-8c24-4b10-9116-83975ed9155a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051368 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqm2g\" (UniqueName: \"kubernetes.io/projected/ee78c68a-a580-49cf-85d4-e172ac11fbab-kube-api-access-tqm2g\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051396 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051416 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4733c1c8-0664-4ccd-b53c-890be583a891-signing-key\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051435 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cft8b\" (UniqueName: \"kubernetes.io/projected/6027adf5-7232-462c-9371-47187637c34c-kube-api-access-cft8b\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051457 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c70994e1-07d8-4989-9483-4dae00ee88a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.051739 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.052251 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b931aa03-f254-4fa4-8f2f-04b735ae01f5-tmpfs\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.052395 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-images\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.052500 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-registration-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.053147 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537827c-6019-4b2d-a15b-8a4d742b1664-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.053545 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee78c68a-a580-49cf-85d4-e172ac11fbab-config\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.054605 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.055271 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36c453df-4dc2-47d5-b66f-688c55b01763-trusted-ca\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.056830 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-mountpoint-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.057735 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c70994e1-07d8-4989-9483-4dae00ee88a9-config\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.058450 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4733c1c8-0664-4ccd-b53c-890be583a891-signing-cabundle\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.061387 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee78c68a-a580-49cf-85d4-e172ac11fbab-serving-cert\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.061791 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ce82c4-8c24-4b10-9116-83975ed9155a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.062707 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-config\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.063636 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-plugins-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.064407 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-default-certificate\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.064457 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-node-bootstrap-token\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.064533 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.065171 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/36c453df-4dc2-47d5-b66f-688c55b01763-metrics-tls\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.065183 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-cert\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.066509 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6027adf5-7232-462c-9371-47187637c34c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.066587 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-socket-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.067506 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-config-volume\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.068393 4888 request.go:700] Waited for 1.86631637s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.070006 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-client\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.070270 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-webhook-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.070568 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f4af2053-39df-4362-aacb-0cd7bc018d9c-csi-data-dir\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.070911 4888 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.071007 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-metrics-certs\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.071012 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fccb657c-29dd-4ccf-a16c-6867e2fd2145-service-ca-bundle\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.071323 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.571306565 +0000 UTC m=+142.153990609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.072431 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2bd354c8-cd5f-43a6-880f-694516d453f7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.072833 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1f8e203-e571-46db-9414-710646e64525-certs\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.073066 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71e1aa4d-7bb1-4621-aada-6fbd776fe776-etcd-service-ca\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.073131 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d8003a3-3839-40a4-a432-1addc1dc99ee-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.073568 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6d39688c-5aad-461d-aa95-d13f144526a7-metrics-tls\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.076530 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.076927 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-metrics-tls\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.077318 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ce82c4-8c24-4b10-9116-83975ed9155a-config\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.077543 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.077754 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.077840 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2bd354c8-cd5f-43a6-880f-694516d453f7-proxy-tls\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.078444 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c70994e1-07d8-4989-9483-4dae00ee88a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.078484 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4733c1c8-0664-4ccd-b53c-890be583a891-signing-key\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.078953 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537827c-6019-4b2d-a15b-8a4d742b1664-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.081234 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/814a1313-1c15-436c-9a1d-132859fabdea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.083400 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.085516 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71e1aa4d-7bb1-4621-aada-6fbd776fe776-serving-cert\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.085737 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-srv-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.086031 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d8003a3-3839-40a4-a432-1addc1dc99ee-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.087132 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.087359 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b931aa03-f254-4fa4-8f2f-04b735ae01f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.088777 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fccb657c-29dd-4ccf-a16c-6867e2fd2145-stats-auth\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.088905 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/81eba091-1a6d-4dff-9b38-33145da7c8bd-srv-cert\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.089511 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.090321 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6027adf5-7232-462c-9371-47187637c34c-proxy-tls\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.091747 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d8111bb-3da5-4474-aee1-efa4173d83ba-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.139671 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.148359 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.152512 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.153284 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.653255624 +0000 UTC m=+142.235939668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.168146 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvlgx\" (UniqueName: \"kubernetes.io/projected/94c7a44d-4411-4588-a723-69464372ce7d-kube-api-access-fvlgx\") pod \"authentication-operator-69f744f599-xz576\" (UID: \"94c7a44d-4411-4588-a723-69464372ce7d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.181203 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/483c2bb0-e41c-4ba2-9644-60781a0ce8f1-kube-api-access-h2rrx\") pod \"cluster-image-registry-operator-dc59b4c8b-jrqpj\" (UID: \"483c2bb0-e41c-4ba2-9644-60781a0ce8f1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.199261 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpwgv\" (UniqueName: \"kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv\") pod \"image-pruner-29399040-nzntm\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.218741 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q26kt"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.229552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg2h4\" (UniqueName: \"kubernetes.io/projected/e71b63a8-b89b-4e4f-83a2-4fad8694e177-kube-api-access-rg2h4\") pod \"openshift-apiserver-operator-796bbdcf4f-6zhf7\" (UID: \"e71b63a8-b89b-4e4f-83a2-4fad8694e177\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:19 crc kubenswrapper[4888]: W1124 00:27:19.235629 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48e64e1d_4554_4633_a7be_970573d5ad06.slice/crio-69f0ea8de4ea436dff2befcb2c9c5b5bb28dd7f7db21d1d47b9724133235c64c WatchSource:0}: Error finding container 69f0ea8de4ea436dff2befcb2c9c5b5bb28dd7f7db21d1d47b9724133235c64c: Status 404 returned error can't find the container with id 69f0ea8de4ea436dff2befcb2c9c5b5bb28dd7f7db21d1d47b9724133235c64c Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.251181 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-png9r\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.255329 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.255422 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.256178 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.756164523 +0000 UTC m=+142.338848567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.257121 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.259293 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.280440 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmmll\" (UniqueName: \"kubernetes.io/projected/103c3930-7610-4c38-af13-cb0f06da2629-kube-api-access-lmmll\") pod \"machine-api-operator-5694c8668f-s8s79\" (UID: \"103c3930-7610-4c38-af13-cb0f06da2629\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.285556 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.297307 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj287\" (UniqueName: \"kubernetes.io/projected/7f7eff25-428d-4192-a4db-2ba3145969a3-kube-api-access-lj287\") pod \"openshift-config-operator-7777fb866f-pqhgl\" (UID: \"7f7eff25-428d-4192-a4db-2ba3145969a3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.302014 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.312142 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:19 crc kubenswrapper[4888]: W1124 00:27:19.313046 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0af5b51e_2144_440f_8d93_dce288248a98.slice/crio-ca6cb5ea4c6797f963ded47a04f6eb7e6a783ddd1a4635f446032f17e43891f2 WatchSource:0}: Error finding container ca6cb5ea4c6797f963ded47a04f6eb7e6a783ddd1a4635f446032f17e43891f2: Status 404 returned error can't find the container with id ca6cb5ea4c6797f963ded47a04f6eb7e6a783ddd1a4635f446032f17e43891f2 Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.319925 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rf5x\" (UniqueName: \"kubernetes.io/projected/34a5f48a-3697-4b07-8922-7007d5a888e2-kube-api-access-8rf5x\") pod \"console-operator-58897d9998-sl4qh\" (UID: \"34a5f48a-3697-4b07-8922-7007d5a888e2\") " pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.338551 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4c7c\" (UniqueName: \"kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c\") pod \"controller-manager-879f6c89f-ff8k7\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.347023 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.357105 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.357711 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.857689918 +0000 UTC m=+142.440373962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.358382 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhh6t\" (UniqueName: \"kubernetes.io/projected/11833290-48d3-4875-a9a6-92b4c9a2a0c2-kube-api-access-hhh6t\") pod \"machine-approver-56656f9798-27bt4\" (UID: \"11833290-48d3-4875-a9a6-92b4c9a2a0c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.382556 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6cfb\" (UniqueName: \"kubernetes.io/projected/e132c48c-e23e-4b5d-bfa6-5858ea0cbeed-kube-api-access-r6cfb\") pod \"cluster-samples-operator-665b6dd947-l8rx2\" (UID: \"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.382997 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.398692 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.404084 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsflc\" (UniqueName: \"kubernetes.io/projected/bdc3d691-d7f9-4dd2-939e-5fc410ddb52c-kube-api-access-bsflc\") pod \"apiserver-7bbb656c7d-cqxhg\" (UID: \"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.418425 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcnf5\" (UniqueName: \"kubernetes.io/projected/81eba091-1a6d-4dff-9b38-33145da7c8bd-kube-api-access-rcnf5\") pod \"olm-operator-6b444d44fb-w249m\" (UID: \"81eba091-1a6d-4dff-9b38-33145da7c8bd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.419354 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.445019 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c70994e1-07d8-4989-9483-4dae00ee88a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2bkj6\" (UID: \"c70994e1-07d8-4989-9483-4dae00ee88a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.459498 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ff2q\" (UniqueName: \"kubernetes.io/projected/71e1aa4d-7bb1-4621-aada-6fbd776fe776-kube-api-access-4ff2q\") pod \"etcd-operator-b45778765-88dzb\" (UID: \"71e1aa4d-7bb1-4621-aada-6fbd776fe776\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.460241 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.460607 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:19.960592987 +0000 UTC m=+142.543277031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.469047 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.482743 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ns44\" (UniqueName: \"kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44\") pod \"marketplace-operator-79b997595-7vpcc\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.489231 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.508768 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.510645 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hndpl\" (UniqueName: \"kubernetes.io/projected/b931aa03-f254-4fa4-8f2f-04b735ae01f5-kube-api-access-hndpl\") pod \"packageserver-d55dfcdfc-4pqlg\" (UID: \"b931aa03-f254-4fa4-8f2f-04b735ae01f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.534948 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r5nx\" (UniqueName: \"kubernetes.io/projected/f4af2053-39df-4362-aacb-0cd7bc018d9c-kube-api-access-4r5nx\") pod \"csi-hostpathplugin-kstd5\" (UID: \"f4af2053-39df-4362-aacb-0cd7bc018d9c\") " pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.545902 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.546912 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5f942"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.553718 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29399040-nzntm"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.553795 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.560570 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.562305 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.562384 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.062363149 +0000 UTC m=+142.645047183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.562717 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.563197 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.063186726 +0000 UTC m=+142.645870780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.565412 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv8g2\" (UniqueName: \"kubernetes.io/projected/5e5d0b2e-798d-4b6d-a012-bdbaaab74b25-kube-api-access-qv8g2\") pod \"dns-default-b52cm\" (UID: \"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25\") " pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.571345 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n5xw\" (UniqueName: \"kubernetes.io/projected/fccb657c-29dd-4ccf-a16c-6867e2fd2145-kube-api-access-4n5xw\") pod \"router-default-5444994796-htnbg\" (UID: \"fccb657c-29dd-4ccf-a16c-6867e2fd2145\") " pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.585928 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsstn\" (UniqueName: \"kubernetes.io/projected/c1f8e203-e571-46db-9414-710646e64525-kube-api-access-tsstn\") pod \"machine-config-server-l99d8\" (UID: \"c1f8e203-e571-46db-9414-710646e64525\") " pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: W1124 00:27:19.593715 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddec11a6c_0eb1_428a_b624_3558ffc42e1b.slice/crio-7fd8fe40ae7452141f10947da5a146fe6354f6d2611660f7c951bca1649d28a9 WatchSource:0}: Error finding container 7fd8fe40ae7452141f10947da5a146fe6354f6d2611660f7c951bca1649d28a9: Status 404 returned error can't find the container with id 7fd8fe40ae7452141f10947da5a146fe6354f6d2611660f7c951bca1649d28a9 Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.595085 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.610689 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4tc4\" (UniqueName: \"kubernetes.io/projected/6d39688c-5aad-461d-aa95-d13f144526a7-kube-api-access-h4tc4\") pod \"dns-operator-744455d44c-ppmph\" (UID: \"6d39688c-5aad-461d-aa95-d13f144526a7\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.612934 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.620440 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl"] Nov 24 00:27:19 crc kubenswrapper[4888]: W1124 00:27:19.627311 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483c2bb0_e41c_4ba2_9644_60781a0ce8f1.slice/crio-1f971c98dbf276781f4bdea1db35ebd5eab622706badf5b6fb340f2771d3038d WatchSource:0}: Error finding container 1f971c98dbf276781f4bdea1db35ebd5eab622706badf5b6fb340f2771d3038d: Status 404 returned error can't find the container with id 1f971c98dbf276781f4bdea1db35ebd5eab622706badf5b6fb340f2771d3038d Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.641668 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjkmz\" (UniqueName: \"kubernetes.io/projected/2bd354c8-cd5f-43a6-880f-694516d453f7-kube-api-access-cjkmz\") pod \"machine-config-operator-74547568cd-j422v\" (UID: \"2bd354c8-cd5f-43a6-880f-694516d453f7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.644578 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5hfj\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-kube-api-access-g5hfj\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.649593 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.662357 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.663452 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.664051 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.164029628 +0000 UTC m=+142.746713672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.665124 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.667254 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk498\" (UniqueName: \"kubernetes.io/projected/a876c9db-47a2-40cf-b004-a4bb7f4b01fa-kube-api-access-dk498\") pod \"catalog-operator-68c6474976-n4brv\" (UID: \"a876c9db-47a2-40cf-b004-a4bb7f4b01fa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.676226 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.681155 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkdmj\" (UniqueName: \"kubernetes.io/projected/3d8003a3-3839-40a4-a432-1addc1dc99ee-kube-api-access-wkdmj\") pod \"openshift-controller-manager-operator-756b6f6bc6-9c2dz\" (UID: \"3d8003a3-3839-40a4-a432-1addc1dc99ee\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.691182 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.699262 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xngd\" (UniqueName: \"kubernetes.io/projected/4733c1c8-0664-4ccd-b53c-890be583a891-kube-api-access-4xngd\") pod \"service-ca-9c57cc56f-c29ws\" (UID: \"4733c1c8-0664-4ccd-b53c-890be583a891\") " pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.706720 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.716680 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xz576"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.721312 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.734081 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36c453df-4dc2-47d5-b66f-688c55b01763-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2rw5m\" (UID: \"36c453df-4dc2-47d5-b66f-688c55b01763\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.736446 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.744534 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.761450 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5vj\" (UniqueName: \"kubernetes.io/projected/42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab-kube-api-access-7b5vj\") pod \"ingress-canary-tpmqr\" (UID: \"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab\") " pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.761612 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x42f\" (UniqueName: \"kubernetes.io/projected/e537827c-6019-4b2d-a15b-8a4d742b1664-kube-api-access-6x42f\") pod \"kube-storage-version-migrator-operator-b67b599dd-2qqgp\" (UID: \"e537827c-6019-4b2d-a15b-8a4d742b1664\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.765883 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.766469 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.266439701 +0000 UTC m=+142.849123955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: W1124 00:27:19.773360 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94c7a44d_4411_4588_a723_69464372ce7d.slice/crio-51074b6919dd33ef686ec94cf367bd3d823d9004e3e6c965dcd5c5053d58a2dc WatchSource:0}: Error finding container 51074b6919dd33ef686ec94cf367bd3d823d9004e3e6c965dcd5c5053d58a2dc: Status 404 returned error can't find the container with id 51074b6919dd33ef686ec94cf367bd3d823d9004e3e6c965dcd5c5053d58a2dc Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.776561 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.781437 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b82bk\" (UniqueName: \"kubernetes.io/projected/00633990-d9ad-48c6-ba3e-dafa1a6bd6ac-kube-api-access-b82bk\") pod \"multus-admission-controller-857f4d67dd-xfrjt\" (UID: \"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.787191 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8s79"] Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.796497 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tpmqr" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.800481 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dznjq\" (UniqueName: \"kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq\") pod \"collect-profiles-29399055-tn92v\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.819174 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l99d8" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.820087 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz7n5\" (UniqueName: \"kubernetes.io/projected/1d8111bb-3da5-4474-aee1-efa4173d83ba-kube-api-access-sz7n5\") pod \"package-server-manager-789f6589d5-b7jtk\" (UID: \"1d8111bb-3da5-4474-aee1-efa4173d83ba\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.832461 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.840056 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2ce82c4-8c24-4b10-9116-83975ed9155a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kvcrd\" (UID: \"a2ce82c4-8c24-4b10-9116-83975ed9155a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.858790 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlzqf\" (UniqueName: \"kubernetes.io/projected/8a357524-1674-4bc5-9f2f-e310387ca2be-kube-api-access-tlzqf\") pod \"migrator-59844c95c7-cldc2\" (UID: \"8a357524-1674-4bc5-9f2f-e310387ca2be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.867077 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.867411 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.367390887 +0000 UTC m=+142.950074931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.887832 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt7fm\" (UniqueName: \"kubernetes.io/projected/814a1313-1c15-436c-9a1d-132859fabdea-kube-api-access-nt7fm\") pod \"control-plane-machine-set-operator-78cbb6b69f-j6829\" (UID: \"814a1313-1c15-436c-9a1d-132859fabdea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.906209 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cft8b\" (UniqueName: \"kubernetes.io/projected/6027adf5-7232-462c-9371-47187637c34c-kube-api-access-cft8b\") pod \"machine-config-controller-84d6567774-ng8n9\" (UID: \"6027adf5-7232-462c-9371-47187637c34c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.924747 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqm2g\" (UniqueName: \"kubernetes.io/projected/ee78c68a-a580-49cf-85d4-e172ac11fbab-kube-api-access-tqm2g\") pod \"service-ca-operator-777779d784-gf7fv\" (UID: \"ee78c68a-a580-49cf-85d4-e172ac11fbab\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.945345 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.971476 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:19 crc kubenswrapper[4888]: E1124 00:27:19.971904 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.471890598 +0000 UTC m=+143.054574642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.981586 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" Nov 24 00:27:19 crc kubenswrapper[4888]: I1124 00:27:19.998911 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.013791 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.028593 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.043748 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.051751 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.058942 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.068328 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88dzb"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.073010 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.073531 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.573508206 +0000 UTC m=+143.156192250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.078195 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sl4qh"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.087184 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.104480 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.111690 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.126231 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5f942" event={"ID":"dec11a6c-0eb1-428a-b624-3558ffc42e1b","Type":"ContainerStarted","Data":"7fd8fe40ae7452141f10947da5a146fe6354f6d2611660f7c951bca1649d28a9"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.131040 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" event={"ID":"e71b63a8-b89b-4e4f-83a2-4fad8694e177","Type":"ContainerStarted","Data":"d7e2867a55ac6441204003f33a5d901deffe8d647ef7184d42fa76c1e2a588c2"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.135178 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29399040-nzntm" event={"ID":"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b","Type":"ContainerStarted","Data":"1702902d58e3a5667e66c2ed60c75eb4e46ac522ac67ed8978bb6d811c17e054"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.161259 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" event={"ID":"11833290-48d3-4875-a9a6-92b4c9a2a0c2","Type":"ContainerStarted","Data":"69a03764e5dcb38156873b15a5fbe4925f9d88d3f6006c7f9a258ba7c90df6da"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.161326 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" event={"ID":"11833290-48d3-4875-a9a6-92b4c9a2a0c2","Type":"ContainerStarted","Data":"364d00f6a21ffe6a5262713522928462a196db179a37bca2d5aeeb0a2467c124"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.174829 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.175581 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.675553637 +0000 UTC m=+143.258237681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.186826 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" event={"ID":"0af5b51e-2144-440f-8d93-dce288248a98","Type":"ContainerStarted","Data":"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.186878 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" event={"ID":"0af5b51e-2144-440f-8d93-dce288248a98","Type":"ContainerStarted","Data":"ca6cb5ea4c6797f963ded47a04f6eb7e6a783ddd1a4635f446032f17e43891f2"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.187788 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.195240 4888 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-mc99g container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.195322 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" podUID="0af5b51e-2144-440f-8d93-dce288248a98" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.196786 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jdqr8" event={"ID":"e594a1de-76b4-4c46-8244-8188ece0c088","Type":"ContainerStarted","Data":"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.196867 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jdqr8" event={"ID":"e594a1de-76b4-4c46-8244-8188ece0c088","Type":"ContainerStarted","Data":"cffbe386c1fde27a96e1e870208c1e9b644617a5485906e999245fedf1c660cf"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.197761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" event={"ID":"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e","Type":"ContainerStarted","Data":"91652a541ff64153e491acc5f12d60583cebd6a4a03838a982257a6526fa77ef"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.199672 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" event={"ID":"94c7a44d-4411-4588-a723-69464372ce7d","Type":"ContainerStarted","Data":"51074b6919dd33ef686ec94cf367bd3d823d9004e3e6c965dcd5c5053d58a2dc"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.201936 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" event={"ID":"b847db54-4cfd-4b84-b3a8-a8aad4304919","Type":"ContainerStarted","Data":"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.201968 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" event={"ID":"b847db54-4cfd-4b84-b3a8-a8aad4304919","Type":"ContainerStarted","Data":"51b68b26ea86aad561dd9cc75c99a4f20733e8363173e6e7e24f84da05b82d86"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.202705 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.204029 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" event={"ID":"483c2bb0-e41c-4ba2-9644-60781a0ce8f1","Type":"ContainerStarted","Data":"1f971c98dbf276781f4bdea1db35ebd5eab622706badf5b6fb340f2771d3038d"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.205644 4888 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-st95l container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.205694 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.207759 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" event={"ID":"48e64e1d-4554-4633-a7be-970573d5ad06","Type":"ContainerDied","Data":"09110fdbab0f2c87ef4c2472576b0efe26f8eef1f414084c325641a9f9e56186"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.211497 4888 generic.go:334] "Generic (PLEG): container finished" podID="48e64e1d-4554-4633-a7be-970573d5ad06" containerID="09110fdbab0f2c87ef4c2472576b0efe26f8eef1f414084c325641a9f9e56186" exitCode=0 Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.211701 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" event={"ID":"48e64e1d-4554-4633-a7be-970573d5ad06","Type":"ContainerStarted","Data":"69f0ea8de4ea436dff2befcb2c9c5b5bb28dd7f7db21d1d47b9724133235c64c"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.222549 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" event={"ID":"7f7eff25-428d-4192-a4db-2ba3145969a3","Type":"ContainerStarted","Data":"de75bb91f8ecc54e7f6a452cc6a80db8adc89066bfdbe4062131a968fa4b6816"} Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.226618 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" event={"ID":"103c3930-7610-4c38-af13-cb0f06da2629","Type":"ContainerStarted","Data":"addbab311e0a0549d216b116a7113fb6ee78c546427a0cdfc52d4d24e1baf276"} Nov 24 00:27:20 crc kubenswrapper[4888]: W1124 00:27:20.240561 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34a5f48a_3697_4b07_8922_7007d5a888e2.slice/crio-2bc62a6e08e8d05438d5be4cbc8f4fb3a725e2fa2f99d1e9378d148b86459279 WatchSource:0}: Error finding container 2bc62a6e08e8d05438d5be4cbc8f4fb3a725e2fa2f99d1e9378d148b86459279: Status 404 returned error can't find the container with id 2bc62a6e08e8d05438d5be4cbc8f4fb3a725e2fa2f99d1e9378d148b86459279 Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.276105 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.276282 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.776252925 +0000 UTC m=+143.358936959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.276560 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.278100 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.778079874 +0000 UTC m=+143.360764108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.365139 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2"] Nov 24 00:27:20 crc kubenswrapper[4888]: W1124 00:27:20.371441 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfccb657c_29dd_4ccf_a16c_6867e2fd2145.slice/crio-877e12c73b94b0130c42bd726f6acaeabf89a1a3769867ddc628594623f4c05a WatchSource:0}: Error finding container 877e12c73b94b0130c42bd726f6acaeabf89a1a3769867ddc628594623f4c05a: Status 404 returned error can't find the container with id 877e12c73b94b0130c42bd726f6acaeabf89a1a3769867ddc628594623f4c05a Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.376049 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.377551 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.379280 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.879256577 +0000 UTC m=+143.461940621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.384919 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kstd5"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.480521 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.484162 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:20.984143441 +0000 UTC m=+143.566827485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.563440 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.564531 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmph"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.581422 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.582116 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.082093659 +0000 UTC m=+143.664777703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.583355 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.700702 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.703612 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.704442 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.204391398 +0000 UTC m=+143.787075452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.705647 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c29ws"] Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.805175 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.805578 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.30555286 +0000 UTC m=+143.888236904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:20 crc kubenswrapper[4888]: I1124 00:27:20.906579 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:20 crc kubenswrapper[4888]: E1124 00:27:20.907211 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.407192059 +0000 UTC m=+143.989876103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.007286 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.008045 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.5080233 +0000 UTC m=+144.090707344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.112889 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.113308 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.613293306 +0000 UTC m=+144.195977350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.195074 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.214889 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.215180 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.715150112 +0000 UTC m=+144.297834156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.215600 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.217042 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.717027403 +0000 UTC m=+144.299711447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.230664 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.234983 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.237204 4888 generic.go:334] "Generic (PLEG): container finished" podID="7f7eff25-428d-4192-a4db-2ba3145969a3" containerID="b8cb4e66da2a330f51d42d526a1802916b612dbb08c87fe9983b3f8a4e7471d7" exitCode=0 Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.237338 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" event={"ID":"7f7eff25-428d-4192-a4db-2ba3145969a3","Type":"ContainerDied","Data":"b8cb4e66da2a330f51d42d526a1802916b612dbb08c87fe9983b3f8a4e7471d7"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.238470 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.239184 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" event={"ID":"6d39688c-5aad-461d-aa95-d13f144526a7","Type":"ContainerStarted","Data":"81acc3b5ac6e2fdf5f7dbae3eaab148e7378939332d504254ae57540d3a5c343"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.242281 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5f942" event={"ID":"dec11a6c-0eb1-428a-b624-3558ffc42e1b","Type":"ContainerStarted","Data":"b721e989990f6937857302bf49b0b6026384cef1940a79a290f6ec4e77d76490"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.242647 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.245013 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" event={"ID":"c70994e1-07d8-4989-9483-4dae00ee88a9","Type":"ContainerStarted","Data":"726670f2c8005c33ba90db1df0c8d33c4902d44013fb8184c3b96f605b335af3"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.259913 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.259986 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.270307 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" event={"ID":"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed","Type":"ContainerStarted","Data":"d547b6d42a73dd137f8b823336184ad6762a7caae46c28ed17a541f13c248711"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.270397 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" podStartSLOduration=123.270370424 podStartE2EDuration="2m3.270370424s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:21.269201956 +0000 UTC m=+143.851886000" watchObservedRunningTime="2025-11-24 00:27:21.270370424 +0000 UTC m=+143.853054478" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.295141 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" event={"ID":"71e1aa4d-7bb1-4621-aada-6fbd776fe776","Type":"ContainerStarted","Data":"936a902145e611f7cf2c1940d0062b591e66dc752f53e46943390239a1bb81c8"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.299156 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-jdqr8" podStartSLOduration=123.299127087 podStartE2EDuration="2m3.299127087s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:21.298790406 +0000 UTC m=+143.881474450" watchObservedRunningTime="2025-11-24 00:27:21.299127087 +0000 UTC m=+143.881811131" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.304532 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" event={"ID":"34a5f48a-3697-4b07-8922-7007d5a888e2","Type":"ContainerStarted","Data":"dc37324429708b2e10bb276862e84608b7108c5e7540dea0915b16440faae72a"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.304599 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" event={"ID":"34a5f48a-3697-4b07-8922-7007d5a888e2","Type":"ContainerStarted","Data":"2bc62a6e08e8d05438d5be4cbc8f4fb3a725e2fa2f99d1e9378d148b86459279"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.305640 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.313962 4888 patch_prober.go:28] interesting pod/console-operator-58897d9998-sl4qh container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.314047 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" podUID="34a5f48a-3697-4b07-8922-7007d5a888e2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.317806 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.320151 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.820102467 +0000 UTC m=+144.402786521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.328096 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.328530 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.82851021 +0000 UTC m=+144.411194254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.324198 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.337086 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j422v"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.353829 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tpmqr"] Nov 24 00:27:21 crc kubenswrapper[4888]: W1124 00:27:21.424869 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod859fc02a_e8f2_4aec_99ce_d00f153ac581.slice/crio-4e7976bfa5fe59cd85a2e6cbd41694eee02e46ebaf9212efd477f5c739000a75 WatchSource:0}: Error finding container 4e7976bfa5fe59cd85a2e6cbd41694eee02e46ebaf9212efd477f5c739000a75: Status 404 returned error can't find the container with id 4e7976bfa5fe59cd85a2e6cbd41694eee02e46ebaf9212efd477f5c739000a75 Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.440381 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b52cm"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.445593 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.448441 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:21.948414501 +0000 UTC m=+144.531098545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.500969 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.513387 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.513454 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.525822 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" event={"ID":"e71b63a8-b89b-4e4f-83a2-4fad8694e177","Type":"ContainerStarted","Data":"8b6eb0d6c1b6d8a5f84596903d9f12196977d0caa675d8738e2980af3fe97e51"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.530551 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.545533 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" event={"ID":"483c2bb0-e41c-4ba2-9644-60781a0ce8f1","Type":"ContainerStarted","Data":"fe63ff2d4f49e7c1f4964922c7ca69cbed884001483d32924dca4c1550778d80"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.550648 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.552936 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.051096843 +0000 UTC m=+144.633781067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.555046 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" event={"ID":"f4af2053-39df-4362-aacb-0cd7bc018d9c","Type":"ContainerStarted","Data":"d911bc13874a3b80a8a263498ec2cb92beb0736721f231b7b6cdde05a4bc54f6"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.558802 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" event={"ID":"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c","Type":"ContainerStarted","Data":"fdca668602a5a48b956a96c59fe0ec0814fa76f41094bdbd53098ebb774cab0c"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.563519 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.568400 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29399040-nzntm" event={"ID":"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b","Type":"ContainerStarted","Data":"c92f8035bd03b463a08c6d01e451cf458d9cd4d65924dfea28ee5e717052e485"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.573822 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.583420 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" podStartSLOduration=123.58339831 podStartE2EDuration="2m3.58339831s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:21.571811284 +0000 UTC m=+144.154495328" watchObservedRunningTime="2025-11-24 00:27:21.58339831 +0000 UTC m=+144.166082354" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.586066 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" event={"ID":"cbb02b82-c40c-4e4b-99dc-7cccd4bd092e","Type":"ContainerStarted","Data":"ab9ff29f39ec05249fa7e63b471121f8295e3387f2b9c47a088a0d3a7f43c0d9"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.604506 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l99d8" event={"ID":"c1f8e203-e571-46db-9414-710646e64525","Type":"ContainerStarted","Data":"8beb16d1441fc5d9d29c88d9925d5d5ea13c2a449eced44817ca619da31811c8"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.604595 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l99d8" event={"ID":"c1f8e203-e571-46db-9414-710646e64525","Type":"ContainerStarted","Data":"73e041885ae67c537827ab83c24bcb21efffa6be9d6faa9ac56dbca7b590508b"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.614155 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.617683 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" event={"ID":"4733c1c8-0664-4ccd-b53c-890be583a891","Type":"ContainerStarted","Data":"6768631c8eb4b25bf1b1c388fbbb2e90a609a9b78f4480d99d4eff7f2e946275"} Nov 24 00:27:21 crc kubenswrapper[4888]: W1124 00:27:21.619265 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod153d6369_b587_4da4_863c_bc8a85edde5e.slice/crio-805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465 WatchSource:0}: Error finding container 805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465: Status 404 returned error can't find the container with id 805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465 Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.634887 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" event={"ID":"c724c643-a37d-41a2-800c-22f0c308f331","Type":"ContainerStarted","Data":"38020d5dc456968cbd2bc2da7ba1ed57815c415bb13257e40eeaed5084eabb9c"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.645905 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" event={"ID":"a876c9db-47a2-40cf-b004-a4bb7f4b01fa","Type":"ContainerStarted","Data":"bf3482087981a4731e0fab7fca7c523e2f4634b79d1c38808ea8aecc25b1f8d7"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.660016 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xfrjt"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.663470 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.664520 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-htnbg" event={"ID":"fccb657c-29dd-4ccf-a16c-6867e2fd2145","Type":"ContainerStarted","Data":"8e6b37496a629d3c626c7df76dfd096c61776a5597b0e8f4da3eb356038c1b67"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.664590 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-htnbg" event={"ID":"fccb657c-29dd-4ccf-a16c-6867e2fd2145","Type":"ContainerStarted","Data":"877e12c73b94b0130c42bd726f6acaeabf89a1a3769867ddc628594623f4c05a"} Nov 24 00:27:21 crc kubenswrapper[4888]: W1124 00:27:21.666140 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ce82c4_8c24_4b10_9116_83975ed9155a.slice/crio-7092e2be9d6f08ced7fa89b624a1ea10d37add73ab1d17fe87fc18e235606bc8 WatchSource:0}: Error finding container 7092e2be9d6f08ced7fa89b624a1ea10d37add73ab1d17fe87fc18e235606bc8: Status 404 returned error can't find the container with id 7092e2be9d6f08ced7fa89b624a1ea10d37add73ab1d17fe87fc18e235606bc8 Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.667652 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.167631014 +0000 UTC m=+144.750315058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.678600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" event={"ID":"103c3930-7610-4c38-af13-cb0f06da2629","Type":"ContainerStarted","Data":"f2ed940f59dff1bb8ac8aaa693059ef4404ae1930e485db04fb29fea5f89f52e"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.681287 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp"] Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.684027 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" event={"ID":"94c7a44d-4411-4588-a723-69464372ce7d","Type":"ContainerStarted","Data":"1d42d633968ec472001b148e1e40972305046c7235d4c90ada7dc215fd9cb4eb"} Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.690121 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.692972 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.765685 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.774516 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.274493481 +0000 UTC m=+144.857177705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.778168 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.780554 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.780641 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 24 00:27:21 crc kubenswrapper[4888]: W1124 00:27:21.837021 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode537827c_6019_4b2d_a15b_8a4d742b1664.slice/crio-90adfb1ba25557cd3db81c0b50e4a4be41a459adb2ba45d398ba352304625f91 WatchSource:0}: Error finding container 90adfb1ba25557cd3db81c0b50e4a4be41a459adb2ba45d398ba352304625f91: Status 404 returned error can't find the container with id 90adfb1ba25557cd3db81c0b50e4a4be41a459adb2ba45d398ba352304625f91 Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.867332 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.867783 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.367749097 +0000 UTC m=+144.950433311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.868044 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.868618 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.368606525 +0000 UTC m=+144.951290569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.899553 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-l99d8" podStartSLOduration=5.899521308 podStartE2EDuration="5.899521308s" podCreationTimestamp="2025-11-24 00:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:21.893510483 +0000 UTC m=+144.476194527" watchObservedRunningTime="2025-11-24 00:27:21.899521308 +0000 UTC m=+144.482205352" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.937251 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6zhf7" podStartSLOduration=123.937228052 podStartE2EDuration="2m3.937228052s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:21.936171908 +0000 UTC m=+144.518855982" watchObservedRunningTime="2025-11-24 00:27:21.937228052 +0000 UTC m=+144.519912096" Nov 24 00:27:21 crc kubenswrapper[4888]: I1124 00:27:21.974310 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:21 crc kubenswrapper[4888]: E1124 00:27:21.974752 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.474719439 +0000 UTC m=+145.057403483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.024421 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-htnbg" podStartSLOduration=124.02439321 podStartE2EDuration="2m4.02439321s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.023711478 +0000 UTC m=+144.606395532" watchObservedRunningTime="2025-11-24 00:27:22.02439321 +0000 UTC m=+144.607077254" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.065775 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5f942" podStartSLOduration=124.065743082 podStartE2EDuration="2m4.065743082s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.056251814 +0000 UTC m=+144.638935858" watchObservedRunningTime="2025-11-24 00:27:22.065743082 +0000 UTC m=+144.648427126" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.079823 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.088448 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.588428448 +0000 UTC m=+145.171112492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.148139 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jrqpj" podStartSLOduration=124.148116585 podStartE2EDuration="2m4.148116585s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.108726947 +0000 UTC m=+144.691410991" watchObservedRunningTime="2025-11-24 00:27:22.148116585 +0000 UTC m=+144.730800619" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.149173 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-xz576" podStartSLOduration=124.149167619 podStartE2EDuration="2m4.149167619s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.146147671 +0000 UTC m=+144.728831835" watchObservedRunningTime="2025-11-24 00:27:22.149167619 +0000 UTC m=+144.731851663" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.184948 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.185441 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.685416256 +0000 UTC m=+145.268100300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.215331 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29399040-nzntm" podStartSLOduration=124.215307385 podStartE2EDuration="2m4.215307385s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.17815988 +0000 UTC m=+144.760843924" watchObservedRunningTime="2025-11-24 00:27:22.215307385 +0000 UTC m=+144.797991429" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.262223 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" podStartSLOduration=124.262199747 podStartE2EDuration="2m4.262199747s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.216785103 +0000 UTC m=+144.799469147" watchObservedRunningTime="2025-11-24 00:27:22.262199747 +0000 UTC m=+144.844883791" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.292521 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.292998 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.792979446 +0000 UTC m=+145.375663500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.316622 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zj2zq" podStartSLOduration=124.316599712 podStartE2EDuration="2m4.316599712s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.308358165 +0000 UTC m=+144.891042219" watchObservedRunningTime="2025-11-24 00:27:22.316599712 +0000 UTC m=+144.899283756" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.394071 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.394958 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.894935684 +0000 UTC m=+145.477619738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.496364 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.496742 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:22.996729107 +0000 UTC m=+145.579413151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.598720 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.598978 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.098959285 +0000 UTC m=+145.681643329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.600874 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.601291 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.10128333 +0000 UTC m=+145.683967374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.703036 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.703281 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.203245749 +0000 UTC m=+145.785929793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.703384 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.703755 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.203727625 +0000 UTC m=+145.786411669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.791445 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" event={"ID":"ee78c68a-a580-49cf-85d4-e172ac11fbab","Type":"ContainerStarted","Data":"dfa41311f7d8bc2b7a0bc96ad8ef72960d40fed29413a0b191bc03c20978e963"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.801078 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" event={"ID":"6027adf5-7232-462c-9371-47187637c34c","Type":"ContainerStarted","Data":"de021820852ad0de1f5b71eac5abd258682863cd1bc380523bb3b215f7d66202"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.801142 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" event={"ID":"6027adf5-7232-462c-9371-47187637c34c","Type":"ContainerStarted","Data":"e8a2994a7a4b43697130ccb8fdebc4dbd33ced3085304b5242eb7b4c30f611a3"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.804689 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.805209 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.305190657 +0000 UTC m=+145.887874701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.811403 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:22 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:22 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:22 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.811461 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.898370 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" event={"ID":"814a1313-1c15-436c-9a1d-132859fabdea","Type":"ContainerStarted","Data":"42a09879d7e8a16d5ca84249237e3d710fc78b51469cb78a71149f284fc6b04d"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.898430 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" event={"ID":"814a1313-1c15-436c-9a1d-132859fabdea","Type":"ContainerStarted","Data":"33db4b1d35198e5c26b98deaa114efcb3d3a56f915573fb9f0a9406f9d880031"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.904661 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" event={"ID":"a876c9db-47a2-40cf-b004-a4bb7f4b01fa","Type":"ContainerStarted","Data":"f59fef62058ee198fc3208a94fdbb462953831cfc86d57f7d522e08b7eac9abb"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.905702 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.906316 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:22 crc kubenswrapper[4888]: E1124 00:27:22.906692 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.406679 +0000 UTC m=+145.989363044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.911157 4888 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-n4brv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.911237 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" podUID="a876c9db-47a2-40cf-b004-a4bb7f4b01fa" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.923915 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j6829" podStartSLOduration=124.923898769 podStartE2EDuration="2m4.923898769s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.922127562 +0000 UTC m=+145.504811606" watchObservedRunningTime="2025-11-24 00:27:22.923898769 +0000 UTC m=+145.506582813" Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.927095 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" event={"ID":"71e1aa4d-7bb1-4621-aada-6fbd776fe776","Type":"ContainerStarted","Data":"65bb441b16c807ee466ae4f2b0629e7d3d36563530c4e37a81937b899b812a1a"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.941949 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" event={"ID":"8a357524-1674-4bc5-9f2f-e310387ca2be","Type":"ContainerStarted","Data":"04f2ce193a47509529a6999a3ccc4f79b5013d4f7fbba32cadc9cb75d6603581"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.967768 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" event={"ID":"2bd354c8-cd5f-43a6-880f-694516d453f7","Type":"ContainerStarted","Data":"1b8b527f5c28ffbbdc850dc03ee555ec624fad402d1a81306c3791791b0ff3a8"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.968288 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" event={"ID":"2bd354c8-cd5f-43a6-880f-694516d453f7","Type":"ContainerStarted","Data":"f85ca006f95e110eee0ff9dcc150684ec6b78c26d0f4dca8c795936ad0f489a5"} Nov 24 00:27:22 crc kubenswrapper[4888]: I1124 00:27:22.972219 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" podStartSLOduration=124.97201345 podStartE2EDuration="2m4.97201345s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:22.969011963 +0000 UTC m=+145.551696007" watchObservedRunningTime="2025-11-24 00:27:22.97201345 +0000 UTC m=+145.554697494" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.011636 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.511589925 +0000 UTC m=+146.094273969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.011696 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.033095 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-88dzb" podStartSLOduration=125.033058991 podStartE2EDuration="2m5.033058991s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.01544183 +0000 UTC m=+145.598125874" watchObservedRunningTime="2025-11-24 00:27:23.033058991 +0000 UTC m=+145.615743045" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.034042 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.054038 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.554009771 +0000 UTC m=+146.136693815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.113021 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" event={"ID":"103c3930-7610-4c38-af13-cb0f06da2629","Type":"ContainerStarted","Data":"a4c32d06945e6ee90637bcc2fd38768685135f42991af37216c723480433528c"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.134903 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.135267 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.635235777 +0000 UTC m=+146.217919821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.135442 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.135855 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.635847247 +0000 UTC m=+146.218531291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.141585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" event={"ID":"1d8111bb-3da5-4474-aee1-efa4173d83ba","Type":"ContainerStarted","Data":"68b4297aef811d4e3b02be4a42199fab43cba18600b9f53630fee32b2da09ca7"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.141651 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" event={"ID":"1d8111bb-3da5-4474-aee1-efa4173d83ba","Type":"ContainerStarted","Data":"c47d63ee60fe13fb24ec10faa35a035acefe2dd785470ec621a7ecf9bb0725be"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.164257 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8s79" podStartSLOduration=125.164238918 podStartE2EDuration="2m5.164238918s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.163553026 +0000 UTC m=+145.746237070" watchObservedRunningTime="2025-11-24 00:27:23.164238918 +0000 UTC m=+145.746922962" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.170596 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" event={"ID":"11833290-48d3-4875-a9a6-92b4c9a2a0c2","Type":"ContainerStarted","Data":"9b6aeb3cc3bb2fcd6efd1b008bddce38722a7c95ad1f81325feee89a66e46cfa"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.194669 4888 generic.go:334] "Generic (PLEG): container finished" podID="bdc3d691-d7f9-4dd2-939e-5fc410ddb52c" containerID="edcadaf10898ccbfe29fe9a9e7d7675b442c004d341b1671f48a12b8f17200b3" exitCode=0 Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.195167 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" event={"ID":"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c","Type":"ContainerDied","Data":"edcadaf10898ccbfe29fe9a9e7d7675b442c004d341b1671f48a12b8f17200b3"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.223930 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.236435 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" event={"ID":"7f7eff25-428d-4192-a4db-2ba3145969a3","Type":"ContainerStarted","Data":"795711a94a9c789a263dcfd0290dfa20b3dc55b7541164622aa8ce06995d5ade"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.237095 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.237193 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.237957 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.73793881 +0000 UTC m=+146.320622854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.266109 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" event={"ID":"b931aa03-f254-4fa4-8f2f-04b735ae01f5","Type":"ContainerStarted","Data":"457414c705356d1de5d8918d87a94cf2749b95a74a256d9b59578d2d043da0c0"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.266194 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" event={"ID":"b931aa03-f254-4fa4-8f2f-04b735ae01f5","Type":"ContainerStarted","Data":"5be233465dee87287b1b1755f789296cd357f6d4754443887219ba9fb947c0cb"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.269259 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.273372 4888 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4pqlg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.273449 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" podUID="b931aa03-f254-4fa4-8f2f-04b735ae01f5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.305773 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" event={"ID":"48e64e1d-4554-4633-a7be-970573d5ad06","Type":"ContainerStarted","Data":"fdca2f4ca93f3e69943c67b80284e7e79e8305036133b2a4f824b03bfcd8ee8b"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.308536 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-27bt4" podStartSLOduration=125.3085202 podStartE2EDuration="2m5.3085202s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.23732771 +0000 UTC m=+145.820011754" watchObservedRunningTime="2025-11-24 00:27:23.3085202 +0000 UTC m=+145.891204244" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.328695 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" event={"ID":"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed","Type":"ContainerStarted","Data":"db555714fc5026d8eaf7684697c33f308e1c94bcf7f8ad1b9fdfccc8c5b2de6e"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.339297 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" podStartSLOduration=125.339265398 podStartE2EDuration="2m5.339265398s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.335477285 +0000 UTC m=+145.918161339" watchObservedRunningTime="2025-11-24 00:27:23.339265398 +0000 UTC m=+145.921949442" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.348385 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.351664 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.851643459 +0000 UTC m=+146.434327693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.382464 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" event={"ID":"859fc02a-e8f2-4aec-99ce-d00f153ac581","Type":"ContainerStarted","Data":"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.382533 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" event={"ID":"859fc02a-e8f2-4aec-99ce-d00f153ac581","Type":"ContainerStarted","Data":"4e7976bfa5fe59cd85a2e6cbd41694eee02e46ebaf9212efd477f5c739000a75"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.384111 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.389642 4888 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7vpcc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.389707 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.401376 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" event={"ID":"a2ce82c4-8c24-4b10-9116-83975ed9155a","Type":"ContainerStarted","Data":"7092e2be9d6f08ced7fa89b624a1ea10d37add73ab1d17fe87fc18e235606bc8"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.409672 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" podStartSLOduration=125.409650612 podStartE2EDuration="2m5.409650612s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.408125602 +0000 UTC m=+145.990809636" watchObservedRunningTime="2025-11-24 00:27:23.409650612 +0000 UTC m=+145.992334656" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.436021 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" event={"ID":"153d6369-b587-4da4-863c-bc8a85edde5e","Type":"ContainerStarted","Data":"805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.436978 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" podStartSLOduration=125.436955738 podStartE2EDuration="2m5.436955738s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.436168902 +0000 UTC m=+146.018852946" watchObservedRunningTime="2025-11-24 00:27:23.436955738 +0000 UTC m=+146.019639782" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.450556 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.452090 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:23.952065218 +0000 UTC m=+146.534749262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.458382 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tpmqr" event={"ID":"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab","Type":"ContainerStarted","Data":"96068d44897855cd747803c04e5237821e8419bca4632554013f6e92f6f9db96"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.467443 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" event={"ID":"c724c643-a37d-41a2-800c-22f0c308f331","Type":"ContainerStarted","Data":"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.469102 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.475803 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" podStartSLOduration=125.475779778 podStartE2EDuration="2m5.475779778s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.470168165 +0000 UTC m=+146.052852229" watchObservedRunningTime="2025-11-24 00:27:23.475779778 +0000 UTC m=+146.058463832" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.484781 4888 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ff8k7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.489463 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" podUID="c724c643-a37d-41a2-800c-22f0c308f331" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.485801 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b52cm" event={"ID":"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25","Type":"ContainerStarted","Data":"d00a4e2861c1d1a4a8b7b62fa3b19740f10591449663aab272d1902a60de25cb"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.490047 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b52cm" event={"ID":"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25","Type":"ContainerStarted","Data":"1ceca17405ce9f6894e7f7a6185f4cb0aa77497d3212b2c31bcec64de1c5fbf4"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.499917 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" event={"ID":"4733c1c8-0664-4ccd-b53c-890be583a891","Type":"ContainerStarted","Data":"0bbc0b3765c79645709b44eee0a26640c05caf1b787c697d231d45e9f6e0c623"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.510135 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" podStartSLOduration=125.510101331 podStartE2EDuration="2m5.510101331s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.500552441 +0000 UTC m=+146.083236485" watchObservedRunningTime="2025-11-24 00:27:23.510101331 +0000 UTC m=+146.092785375" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.527897 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" podStartSLOduration=125.527875118 podStartE2EDuration="2m5.527875118s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.51560374 +0000 UTC m=+146.098287784" watchObservedRunningTime="2025-11-24 00:27:23.527875118 +0000 UTC m=+146.110559162" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.529770 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" event={"ID":"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac","Type":"ContainerStarted","Data":"1c30e51c83bc27c109bb8341160f48a5b3cad5cabfde497b72397dbcb90f7332"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.532244 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" event={"ID":"c70994e1-07d8-4989-9483-4dae00ee88a9","Type":"ContainerStarted","Data":"c81996874e4a0fa1acadf25fc0eb23e4e05df1bb0afc3e8b7098db7ca3594a65"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.553233 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" event={"ID":"e537827c-6019-4b2d-a15b-8a4d742b1664","Type":"ContainerStarted","Data":"90adfb1ba25557cd3db81c0b50e4a4be41a459adb2ba45d398ba352304625f91"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.557977 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.559807 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.059787504 +0000 UTC m=+146.642471548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.565156 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" event={"ID":"3d8003a3-3839-40a4-a432-1addc1dc99ee","Type":"ContainerStarted","Data":"7cef89c10c17f9c810d4eeab36878eaf4cf70f3b79225cabef1fd171f8b280ea"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.565216 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" event={"ID":"3d8003a3-3839-40a4-a432-1addc1dc99ee","Type":"ContainerStarted","Data":"3a77c504ccc45b3e8cf2859579ad8e6e955428130a2e81b1b482b67eec1ba997"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.576204 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" event={"ID":"36c453df-4dc2-47d5-b66f-688c55b01763","Type":"ContainerStarted","Data":"a62a87e5f981ce44b1551597a40a3e3b30a77d633f33ce3766845dd2679283d9"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.583242 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" event={"ID":"81eba091-1a6d-4dff-9b38-33145da7c8bd","Type":"ContainerStarted","Data":"7515aadef7790e95873ef66e92ec25d7f23a70df8000ec446ef387302bdf515d"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.584269 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.590929 4888 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-w249m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.590993 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" podUID="81eba091-1a6d-4dff-9b38-33145da7c8bd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.591117 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.591139 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.594227 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" podStartSLOduration=125.594210521 podStartE2EDuration="2m5.594210521s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.561084816 +0000 UTC m=+146.143768850" watchObservedRunningTime="2025-11-24 00:27:23.594210521 +0000 UTC m=+146.176894565" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.599793 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" event={"ID":"6d39688c-5aad-461d-aa95-d13f144526a7","Type":"ContainerStarted","Data":"4afab47d69cda2d3cf853b561bcfa221f74dbdc407c3c132d16178197c23b04b"} Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.600569 4888 patch_prober.go:28] interesting pod/console-operator-58897d9998-sl4qh container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.600614 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" podUID="34a5f48a-3697-4b07-8922-7007d5a888e2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.601011 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.601045 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.603323 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-tpmqr" podStartSLOduration=7.603291725 podStartE2EDuration="7.603291725s" podCreationTimestamp="2025-11-24 00:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.59759589 +0000 UTC m=+146.180279934" watchObservedRunningTime="2025-11-24 00:27:23.603291725 +0000 UTC m=+146.185975779" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.626267 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-c29ws" podStartSLOduration=125.62623938 podStartE2EDuration="2m5.62623938s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.621812976 +0000 UTC m=+146.204497020" watchObservedRunningTime="2025-11-24 00:27:23.62623938 +0000 UTC m=+146.208923424" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.663547 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.668574 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.168539862 +0000 UTC m=+146.751224057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.729889 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2bkj6" podStartSLOduration=125.729854072 podStartE2EDuration="2m5.729854072s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.662351722 +0000 UTC m=+146.245035766" watchObservedRunningTime="2025-11-24 00:27:23.729854072 +0000 UTC m=+146.312538116" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.768325 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.806932 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" podStartSLOduration=125.806854121 podStartE2EDuration="2m5.806854121s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.729381207 +0000 UTC m=+146.312065251" watchObservedRunningTime="2025-11-24 00:27:23.806854121 +0000 UTC m=+146.389538185" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.808224 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" podStartSLOduration=125.808217155 podStartE2EDuration="2m5.808217155s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.770062937 +0000 UTC m=+146.352746981" watchObservedRunningTime="2025-11-24 00:27:23.808217155 +0000 UTC m=+146.390901199" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.816963 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:23 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:23 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:23 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.817025 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.818426 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.318410376 +0000 UTC m=+146.901094420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.836769 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9c2dz" podStartSLOduration=125.836741521 podStartE2EDuration="2m5.836741521s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.828878065 +0000 UTC m=+146.411562139" watchObservedRunningTime="2025-11-24 00:27:23.836741521 +0000 UTC m=+146.419425575" Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.872960 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.874136 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" podStartSLOduration=125.874111653 podStartE2EDuration="2m5.874111653s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:23.87246187 +0000 UTC m=+146.455145914" watchObservedRunningTime="2025-11-24 00:27:23.874111653 +0000 UTC m=+146.456795697" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.875295 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.375159797 +0000 UTC m=+146.957843841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:23 crc kubenswrapper[4888]: I1124 00:27:23.976097 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:23 crc kubenswrapper[4888]: E1124 00:27:23.976516 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.476500556 +0000 UTC m=+147.059184600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.077763 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.078036 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.577999419 +0000 UTC m=+147.160683453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.078191 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.078643 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.57863627 +0000 UTC m=+147.161320314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.179466 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.179698 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.679661538 +0000 UTC m=+147.262345582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.180196 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.180663 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.68064271 +0000 UTC m=+147.263326754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.281777 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.281995 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.781962238 +0000 UTC m=+147.364646282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.282278 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.282758 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.782746983 +0000 UTC m=+147.365431227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.383678 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.384085 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.884066531 +0000 UTC m=+147.466750575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.485137 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.485636 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:24.985619887 +0000 UTC m=+147.568303921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.586511 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.587034 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.087012697 +0000 UTC m=+147.669696741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.607022 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmph" event={"ID":"6d39688c-5aad-461d-aa95-d13f144526a7","Type":"ContainerStarted","Data":"f0756bbb07f4bca9f673a51d0bfb7dc065a5658177b9a3b9f4674da309cb0cb6"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.609490 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" event={"ID":"f4af2053-39df-4362-aacb-0cd7bc018d9c","Type":"ContainerStarted","Data":"5d5c524b02406eaabb51771966a156c0b5870501efe33badcd547d670dd481ac"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.610753 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2qqgp" event={"ID":"e537827c-6019-4b2d-a15b-8a4d742b1664","Type":"ContainerStarted","Data":"3be2ace8a20754d8be9aa3ffd96f384a7638e9f2cdae8d34689315730421b2ba"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.612745 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" event={"ID":"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac","Type":"ContainerStarted","Data":"02b99b77c642483392650f347f7a6cf50ba53e461e7d52e10c194b23d6ed3af3"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.612771 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" event={"ID":"00633990-d9ad-48c6-ba3e-dafa1a6bd6ac","Type":"ContainerStarted","Data":"b19cdd3db3f89808faf4a764c135990b1399bb6d49fa82988bea43725f974924"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.616030 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" event={"ID":"48e64e1d-4554-4633-a7be-970573d5ad06","Type":"ContainerStarted","Data":"d07bd37d176fdddda8c2919dddbd215ec11b6f99f72c4100a783bacd4a12b45c"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.618069 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" event={"ID":"2bd354c8-cd5f-43a6-880f-694516d453f7","Type":"ContainerStarted","Data":"351a23ff7f6831d2653a11d78392d4782217e118fd64f92ffe74ce651e662c2a"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.620218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" event={"ID":"1d8111bb-3da5-4474-aee1-efa4173d83ba","Type":"ContainerStarted","Data":"5f85c16879bb0af1afdaa0e1255cd385573216112edbccf57a7de5e3a6bb2ebb"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.620313 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.621892 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" event={"ID":"81eba091-1a6d-4dff-9b38-33145da7c8bd","Type":"ContainerStarted","Data":"847654862b81193e9339e88c51faefab05b2a4399b62786f5d2ef398e786e056"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.622630 4888 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-w249m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.622664 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" podUID="81eba091-1a6d-4dff-9b38-33145da7c8bd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.624412 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-l8rx2" event={"ID":"e132c48c-e23e-4b5d-bfa6-5858ea0cbeed","Type":"ContainerStarted","Data":"f6ae4f471e1c3d90a2290e36702eb2b6b37423e4dc1771d4db27ea21b5aef9aa"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.626340 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" event={"ID":"153d6369-b587-4da4-863c-bc8a85edde5e","Type":"ContainerStarted","Data":"c15a354f669900d9d4d784de64de4fd086f9d80c2f1a5f44ce47ef4031cb240f"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.628541 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" event={"ID":"bdc3d691-d7f9-4dd2-939e-5fc410ddb52c","Type":"ContainerStarted","Data":"e574d877d50750ba7d984fc324290c579b32fdea1e09f7ecf9bc704123ac495a"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.630718 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kvcrd" event={"ID":"a2ce82c4-8c24-4b10-9116-83975ed9155a","Type":"ContainerStarted","Data":"a165c157e7fd00a3ff24eb06284749418af48c3fc038154f5adc4824f4a34561"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.632711 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b52cm" event={"ID":"5e5d0b2e-798d-4b6d-a012-bdbaaab74b25","Type":"ContainerStarted","Data":"bf0b403329da7765fde0e81d899db00f060c74795e0842c99a2ef5b6524029cf"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.633101 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.634613 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" event={"ID":"8a357524-1674-4bc5-9f2f-e310387ca2be","Type":"ContainerStarted","Data":"f5bda9f329ecf735b335dd12f0c3f72bfd3f5e5c0cb802969b7bf4a096054290"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.634635 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" event={"ID":"8a357524-1674-4bc5-9f2f-e310387ca2be","Type":"ContainerStarted","Data":"f01d96ed71f551b090557c6caa62b8ab180564b98bca1f28ceef3321d8d865ab"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.637015 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" event={"ID":"36c453df-4dc2-47d5-b66f-688c55b01763","Type":"ContainerStarted","Data":"dbd3bd45522098b1fd7f6d22ee6e1efebcadd4ac39688cf1462e5b16e86062bc"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.637045 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" event={"ID":"36c453df-4dc2-47d5-b66f-688c55b01763","Type":"ContainerStarted","Data":"84474449b1e97d81202eec39b69a2366068507965cd6b455830d662ef70ecaaf"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.638668 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" event={"ID":"6027adf5-7232-462c-9371-47187637c34c","Type":"ContainerStarted","Data":"20c13c6a0a7789012a29832dd917d83b6624d192bed80b1f76b4186743e9beec"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.640537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" event={"ID":"ee78c68a-a580-49cf-85d4-e172ac11fbab","Type":"ContainerStarted","Data":"d0c742a37208abbc290426ba0f41f39247508b5af5b7fe90f11b0b14a2258d54"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.642313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tpmqr" event={"ID":"42334bb3-ac36-4f1f-9a1e-481ef0dbf2ab","Type":"ContainerStarted","Data":"c55da05e863894d5298157beed50d8870ee29de636703f346da3c8696afe4200"} Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.643676 4888 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7vpcc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.643712 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.644095 4888 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4pqlg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.644120 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" podUID="b931aa03-f254-4fa4-8f2f-04b735ae01f5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.655467 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-xfrjt" podStartSLOduration=126.655447427 podStartE2EDuration="2m6.655447427s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.651595272 +0000 UTC m=+147.234279316" watchObservedRunningTime="2025-11-24 00:27:24.655447427 +0000 UTC m=+147.238131471" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.657004 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4brv" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.696943 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.697143 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.19712717 +0000 UTC m=+147.779811214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.738306 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.743544 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cldc2" podStartSLOduration=126.743458193 podStartE2EDuration="2m6.743458193s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.690482514 +0000 UTC m=+147.273166558" watchObservedRunningTime="2025-11-24 00:27:24.743458193 +0000 UTC m=+147.326142247" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.744409 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" podStartSLOduration=126.744399764 podStartE2EDuration="2m6.744399764s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.739413662 +0000 UTC m=+147.322097706" watchObservedRunningTime="2025-11-24 00:27:24.744399764 +0000 UTC m=+147.327083818" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.783169 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:24 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:24 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:24 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.783234 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.798372 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.799679 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.299655797 +0000 UTC m=+147.882339841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.869794 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" podStartSLOduration=126.869768112 podStartE2EDuration="2m6.869768112s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.866282569 +0000 UTC m=+147.448966623" watchObservedRunningTime="2025-11-24 00:27:24.869768112 +0000 UTC m=+147.452452156" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.900634 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:24 crc kubenswrapper[4888]: E1124 00:27:24.901582 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.401554933 +0000 UTC m=+147.984239147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.904767 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gf7fv" podStartSLOduration=126.904745647 podStartE2EDuration="2m6.904745647s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.903638661 +0000 UTC m=+147.486322705" watchObservedRunningTime="2025-11-24 00:27:24.904745647 +0000 UTC m=+147.487429701" Nov 24 00:27:24 crc kubenswrapper[4888]: I1124 00:27:24.953937 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-b52cm" podStartSLOduration=8.953909262 podStartE2EDuration="8.953909262s" podCreationTimestamp="2025-11-24 00:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.950667157 +0000 UTC m=+147.533351201" watchObservedRunningTime="2025-11-24 00:27:24.953909262 +0000 UTC m=+147.536593306" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.001931 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.002581 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.502544231 +0000 UTC m=+148.085228275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.055121 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j422v" podStartSLOduration=127.055098496 podStartE2EDuration="2m7.055098496s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:25.048613045 +0000 UTC m=+147.631297089" watchObservedRunningTime="2025-11-24 00:27:25.055098496 +0000 UTC m=+147.637782540" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.055664 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2rw5m" podStartSLOduration=127.055658524 podStartE2EDuration="2m7.055658524s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:24.999543183 +0000 UTC m=+147.582227237" watchObservedRunningTime="2025-11-24 00:27:25.055658524 +0000 UTC m=+147.638342568" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.079816 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" podStartSLOduration=127.079781857 podStartE2EDuration="2m7.079781857s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:25.077738931 +0000 UTC m=+147.660422975" watchObservedRunningTime="2025-11-24 00:27:25.079781857 +0000 UTC m=+147.662465901" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.103325 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.103669 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.603654462 +0000 UTC m=+148.186338506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.103968 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.104957 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ng8n9" podStartSLOduration=127.104934713 podStartE2EDuration="2m7.104934713s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:25.102313138 +0000 UTC m=+147.684997182" watchObservedRunningTime="2025-11-24 00:27:25.104934713 +0000 UTC m=+147.687618757" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.116705 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.205865 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.206019 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.705985581 +0000 UTC m=+148.288669625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.206160 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.206196 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.206234 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.206281 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.206515 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.706507188 +0000 UTC m=+148.289191232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.210975 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.213151 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.216477 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.275134 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.309623 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.310015 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.809995736 +0000 UTC m=+148.392679780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.411804 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.412639 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:25.912625756 +0000 UTC m=+148.495309800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.441526 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pqhgl" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.466005 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.476916 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.506229 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.507255 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.519661 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.519981 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.520101 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.020077443 +0000 UTC m=+148.602761487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.595965 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.625753 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.626097 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.126082513 +0000 UTC m=+148.708766557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.626297 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.626322 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4rwb\" (UniqueName: \"kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.626367 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.720214 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.734169 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.727689 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.22766595 +0000 UTC m=+148.810349994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.727589 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.742335 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.742539 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.742642 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4rwb\" (UniqueName: \"kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.742833 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.743871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.744256 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.244232187 +0000 UTC m=+148.826916401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.745095 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.748620 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" event={"ID":"f4af2053-39df-4362-aacb-0cd7bc018d9c","Type":"ContainerStarted","Data":"cc3b1be4614a6088a2bd01a88da7e9f8b1ff74c003c42e328e0e7c59bbb3f8dc"} Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.760343 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.768330 4888 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7vpcc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.768379 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.786549 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.810670 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:25 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:25 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:25 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.810716 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.817633 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w249m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.840028 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4rwb\" (UniqueName: \"kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb\") pod \"certified-operators-ngz8m\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.850598 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.850917 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nksrp\" (UniqueName: \"kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.851083 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.851421 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.852854 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.35280589 +0000 UTC m=+148.935490094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.903950 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.930361 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.931488 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.952523 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.952564 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nksrp\" (UniqueName: \"kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.952622 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.952698 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.953104 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:25 crc kubenswrapper[4888]: E1124 00:27:25.953376 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.453363313 +0000 UTC m=+149.036047357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:25 crc kubenswrapper[4888]: I1124 00:27:25.956269 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.049009 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.055075 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.055486 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.055674 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.055707 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwtq6\" (UniqueName: \"kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.055865 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.555823158 +0000 UTC m=+149.138507202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.150342 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nksrp\" (UniqueName: \"kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp\") pod \"community-operators-mkctm\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.150697 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.162689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.162762 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.162788 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwtq6\" (UniqueName: \"kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.162834 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.163234 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.163511 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.663497772 +0000 UTC m=+149.246181816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.163916 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.164232 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.165454 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.209490 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.243864 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwtq6\" (UniqueName: \"kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6\") pod \"certified-operators-sg6wm\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.267257 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.267426 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.267460 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkmzq\" (UniqueName: \"kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.267499 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.267590 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.767570159 +0000 UTC m=+149.350254203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.340032 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.369652 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.369711 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkmzq\" (UniqueName: \"kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.369755 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.369780 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.370180 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.870166858 +0000 UTC m=+149.452850902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.370766 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.370862 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.387458 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pqlg" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.428747 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkmzq\" (UniqueName: \"kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq\") pod \"community-operators-9hrl6\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.471284 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.471578 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:26.971551878 +0000 UTC m=+149.554235922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.561410 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.584630 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.585000 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.084986189 +0000 UTC m=+149.667670233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.690184 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.690939 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.190921516 +0000 UTC m=+149.773605560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.787047 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:26 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:26 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:26 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.787116 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.793387 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.794002 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.293982721 +0000 UTC m=+149.876666765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.823318 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a3daf031a4cb2b92486fd0cdba255249bc2b16a391ab9ba4d327d94ddf274676"} Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.890218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" event={"ID":"f4af2053-39df-4362-aacb-0cd7bc018d9c","Type":"ContainerStarted","Data":"dee2ea342f8b955a299fed32b92f15efccacb71e749821cc1ea35b8926ee789d"} Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.897588 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:26 crc kubenswrapper[4888]: E1124 00:27:26.898040 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.398019907 +0000 UTC m=+149.980703951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.940681 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b7bb05225d1c47e23a1429732240674a37aea2719f0b898badf328b98a9b6e7c"} Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.966713 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ad0239c237f12800060db5a154d78ad56928a9ed7d911a10cbcc78190e61be8b"} Nov 24 00:27:26 crc kubenswrapper[4888]: I1124 00:27:26.966761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a8f92851d02f6990ffb94e16d86e709e2de0f7e1f7a9d38f471d88ced3207a73"} Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:26.999167 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: E1124 00:27:27.016107 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.516077768 +0000 UTC m=+150.098761992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.075784 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.100330 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:27 crc kubenswrapper[4888]: E1124 00:27:27.100681 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.600661882 +0000 UTC m=+150.183345926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:27 crc kubenswrapper[4888]: W1124 00:27:27.100731 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c084ed8_8f40_4bba_b2d8_07992d26fd35.slice/crio-bfc32be6325894c640a0441a2ba0c44f1e50262b7493137bb7be1150d86a9e1b WatchSource:0}: Error finding container bfc32be6325894c640a0441a2ba0c44f1e50262b7493137bb7be1150d86a9e1b: Status 404 returned error can't find the container with id bfc32be6325894c640a0441a2ba0c44f1e50262b7493137bb7be1150d86a9e1b Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.195693 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.208843 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: E1124 00:27:27.209329 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.709312298 +0000 UTC m=+150.291996342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.301107 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.312065 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:27 crc kubenswrapper[4888]: E1124 00:27:27.312494 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.812469755 +0000 UTC m=+150.395153789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.373068 4888 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.413401 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: E1124 00:27:27.414112 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 00:27:27.914084433 +0000 UTC m=+150.496768477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d8t6" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.461951 4888 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T00:27:27.373097533Z","Handler":null,"Name":""} Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.465582 4888 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.465629 4888 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.525211 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.555360 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.628699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.633816 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.634018 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.656793 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.704161 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d8t6\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.706058 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.708813 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.712677 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.728717 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.729336 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.785022 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:27 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:27 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:27 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.785089 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.830764 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.830877 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.830932 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vch5n\" (UniqueName: \"kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.931920 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vch5n\" (UniqueName: \"kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.932409 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.932486 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.933436 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.934241 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.976286 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vch5n\" (UniqueName: \"kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n\") pod \"redhat-marketplace-7fx6r\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.984831 4888 generic.go:334] "Generic (PLEG): container finished" podID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerID="96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8" exitCode=0 Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.984946 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerDied","Data":"96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8"} Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.984987 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerStarted","Data":"4e4454eb035152a7ba54cfe14508275452580a8b0a6592242c4780203cbca785"} Nov 24 00:27:27 crc kubenswrapper[4888]: I1124 00:27:27.996938 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.009888 4888 generic.go:334] "Generic (PLEG): container finished" podID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerID="a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685" exitCode=0 Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.010208 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerDied","Data":"a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.010358 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerStarted","Data":"bfc32be6325894c640a0441a2ba0c44f1e50262b7493137bb7be1150d86a9e1b"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.023233 4888 generic.go:334] "Generic (PLEG): container finished" podID="153d6369-b587-4da4-863c-bc8a85edde5e" containerID="c15a354f669900d9d4d784de64de4fd086f9d80c2f1a5f44ce47ef4031cb240f" exitCode=0 Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.023302 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" event={"ID":"153d6369-b587-4da4-863c-bc8a85edde5e","Type":"ContainerDied","Data":"c15a354f669900d9d4d784de64de4fd086f9d80c2f1a5f44ce47ef4031cb240f"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.039645 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerStarted","Data":"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.039710 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerStarted","Data":"ed9d76efc3c3c7b33f364019e12ebadd49c6412914c33961d77a6816beaa0d45"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.050561 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" event={"ID":"f4af2053-39df-4362-aacb-0cd7bc018d9c","Type":"ContainerStarted","Data":"a5ff1dea2a0b3171605be263627e13e4efadd3a8db8e49aed3834b74d9a40b44"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.053440 4888 generic.go:334] "Generic (PLEG): container finished" podID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerID="1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4" exitCode=0 Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.053553 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerDied","Data":"1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.053711 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerStarted","Data":"5a48d51d1c393f1bf8ab9b257a5f8d75c324f3b2e0e3aec783864d5ba0e7de22"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.080664 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.082418 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"435cccd5b791221231cbec7064597894dbde0f83844c833ca8764760019a191d"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.083093 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.094689 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.096263 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.108494 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"aa90d62b353a70744e0e9c0545c1865bc1f219d5af248330482759a164e65901"} Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.137695 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.137754 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqks\" (UniqueName: \"kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.137926 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.151510 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.185144 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-kstd5" podStartSLOduration=11.185120513 podStartE2EDuration="11.185120513s" podCreationTimestamp="2025-11-24 00:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:28.184344497 +0000 UTC m=+150.767028541" watchObservedRunningTime="2025-11-24 00:27:28.185120513 +0000 UTC m=+150.767804557" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.202784 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.240658 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.240700 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqks\" (UniqueName: \"kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.240779 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.241185 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.241653 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.274128 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.311409 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqks\" (UniqueName: \"kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks\") pod \"redhat-marketplace-jknzx\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.476919 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.529904 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.539145 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.540966 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.547142 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.547334 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.569678 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.664947 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.665031 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.679834 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.681902 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.689913 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.690365 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.690741 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.692348 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.694048 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.708916 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.717479 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768603 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768680 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768761 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768780 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srsj7\" (UniqueName: \"kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.768816 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.769035 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.769125 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.788657 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:28 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:28 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:28 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.788729 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.811236 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.845443 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.865403 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.870412 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.870511 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.870539 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srsj7\" (UniqueName: \"kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.871247 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.871567 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.871683 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.871719 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.871807 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.891054 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srsj7\" (UniqueName: \"kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7\") pod \"redhat-operators-fgpsq\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.891993 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.979039 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.979089 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:28 crc kubenswrapper[4888]: I1124 00:27:28.998887 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.003640 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.005477 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.010060 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.015482 4888 patch_prober.go:28] interesting pod/console-f9d7485db-jdqr8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.015579 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-jdqr8" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.023121 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.023264 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.023329 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.023335 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.023480 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.095543 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.097030 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.124178 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.136572 4888 generic.go:334] "Generic (PLEG): container finished" podID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerID="cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7" exitCode=0 Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.136634 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerDied","Data":"cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.136665 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerStarted","Data":"7f18008515a37eef341febcd0e01bc031b1e5a2646e1cb6dcf677332ba168deb"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.159109 4888 generic.go:334] "Generic (PLEG): container finished" podID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerID="646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d" exitCode=0 Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.159208 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerDied","Data":"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.175481 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" event={"ID":"1a2cedaf-0125-42ee-969b-59699128ebcc","Type":"ContainerStarted","Data":"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.175551 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" event={"ID":"1a2cedaf-0125-42ee-969b-59699128ebcc","Type":"ContainerStarted","Data":"364d0be16ea1cd06e68ad8e1db0646ca7cf6719bf644023158c5a9bb620c0cda"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.176759 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwxj\" (UniqueName: \"kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.176840 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.176858 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.187243 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerStarted","Data":"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.187295 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerStarted","Data":"86cfe8b83975ceeed94b8c24a6da0749ad38946779cc66360e6608a1bb323f9d"} Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.192033 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q26kt" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.213974 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" podStartSLOduration=131.213941417 podStartE2EDuration="2m11.213941417s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:29.210099912 +0000 UTC m=+151.792783956" watchObservedRunningTime="2025-11-24 00:27:29.213941417 +0000 UTC m=+151.796625471" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.285201 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwxj\" (UniqueName: \"kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.285289 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.285332 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.286809 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.289725 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.360110 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.449780 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwxj\" (UniqueName: \"kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj\") pod \"redhat-operators-hvhgl\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: W1124 00:27:29.467693 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode726333f_6370_454e_8cf8_79a3b8c72227.slice/crio-847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181 WatchSource:0}: Error finding container 847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181: Status 404 returned error can't find the container with id 847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181 Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.497604 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.529695 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sl4qh" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.613683 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.613723 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.636098 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.705706 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.737617 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.739444 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.782329 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.790437 4888 patch_prober.go:28] interesting pod/router-default-5444994796-htnbg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 00:27:29 crc kubenswrapper[4888]: [-]has-synced failed: reason withheld Nov 24 00:27:29 crc kubenswrapper[4888]: [+]process-running ok Nov 24 00:27:29 crc kubenswrapper[4888]: healthz check failed Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.790499 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-htnbg" podUID="fccb657c-29dd-4ccf-a16c-6867e2fd2145" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.794100 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume\") pod \"153d6369-b587-4da4-863c-bc8a85edde5e\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.794164 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume\") pod \"153d6369-b587-4da4-863c-bc8a85edde5e\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.794193 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dznjq\" (UniqueName: \"kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq\") pod \"153d6369-b587-4da4-863c-bc8a85edde5e\" (UID: \"153d6369-b587-4da4-863c-bc8a85edde5e\") " Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.796519 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume" (OuterVolumeSpecName: "config-volume") pod "153d6369-b587-4da4-863c-bc8a85edde5e" (UID: "153d6369-b587-4da4-863c-bc8a85edde5e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.805720 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq" (OuterVolumeSpecName: "kube-api-access-dznjq") pod "153d6369-b587-4da4-863c-bc8a85edde5e" (UID: "153d6369-b587-4da4-863c-bc8a85edde5e"). InnerVolumeSpecName "kube-api-access-dznjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.811051 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "153d6369-b587-4da4-863c-bc8a85edde5e" (UID: "153d6369-b587-4da4-863c-bc8a85edde5e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.822652 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.895787 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153d6369-b587-4da4-863c-bc8a85edde5e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.895835 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dznjq\" (UniqueName: \"kubernetes.io/projected/153d6369-b587-4da4-863c-bc8a85edde5e-kube-api-access-dznjq\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:29 crc kubenswrapper[4888]: I1124 00:27:29.895847 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153d6369-b587-4da4-863c-bc8a85edde5e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.153310 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:27:30 crc kubenswrapper[4888]: W1124 00:27:30.180502 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc54474df_3d22_4c12_a402_47fb8d8ee014.slice/crio-4c3e6e991f2a2d17dd837f06432bb35e16ff1de80e51aaba11f42d3f7572de44 WatchSource:0}: Error finding container 4c3e6e991f2a2d17dd837f06432bb35e16ff1de80e51aaba11f42d3f7572de44: Status 404 returned error can't find the container with id 4c3e6e991f2a2d17dd837f06432bb35e16ff1de80e51aaba11f42d3f7572de44 Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.199484 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" event={"ID":"153d6369-b587-4da4-863c-bc8a85edde5e","Type":"ContainerDied","Data":"805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465"} Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.199518 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805b084231474c5300cdfc2cccee231b45e0b173c4a59d9507afe12463a51465" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.201336 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e726333f-6370-454e-8cf8-79a3b8c72227","Type":"ContainerStarted","Data":"847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181"} Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.201812 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.205984 4888 generic.go:334] "Generic (PLEG): container finished" podID="60936af3-7a00-49b5-8293-626b37cb41e5" containerID="53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14" exitCode=0 Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.206047 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerDied","Data":"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14"} Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.214537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"277e10a8-1c23-4683-9be9-0039e28cc55f","Type":"ContainerStarted","Data":"b8f10a428d379d89290920deb89f705e7cd7127a7d7366a246dbcb1caddd79e0"} Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.231147 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerStarted","Data":"6b3b62bce8efdebb709e1cdd56f84e9b5f84fadcf29a5ce4bd7534ff7498518e"} Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.231612 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.236195 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cqxhg" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.781598 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:30 crc kubenswrapper[4888]: I1124 00:27:30.784233 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-htnbg" Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.266435 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"277e10a8-1c23-4683-9be9-0039e28cc55f","Type":"ContainerStarted","Data":"56eb51323f079c24a64eb9bf89416aa303002a66fb51a45a153b00120fa41b3b"} Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.276800 4888 generic.go:334] "Generic (PLEG): container finished" podID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerID="1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65" exitCode=0 Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.276976 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerDied","Data":"1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65"} Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.281208 4888 generic.go:334] "Generic (PLEG): container finished" podID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerID="f6d0ef01e87f4837f43e15f54e5a8b5bf29a45f516dbdb818afb6a9a73179189" exitCode=0 Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.281298 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerDied","Data":"f6d0ef01e87f4837f43e15f54e5a8b5bf29a45f516dbdb818afb6a9a73179189"} Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.281331 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerStarted","Data":"4c3e6e991f2a2d17dd837f06432bb35e16ff1de80e51aaba11f42d3f7572de44"} Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.288113 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.288099332 podStartE2EDuration="3.288099332s" podCreationTimestamp="2025-11-24 00:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:31.286957905 +0000 UTC m=+153.869641949" watchObservedRunningTime="2025-11-24 00:27:31.288099332 +0000 UTC m=+153.870783376" Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.289849 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e726333f-6370-454e-8cf8-79a3b8c72227","Type":"ContainerDied","Data":"cbce5ece5719b7d88f8d6e16bacce79fd3e307a1604cc4623c49f68e1f724ab5"} Nov 24 00:27:31 crc kubenswrapper[4888]: I1124 00:27:31.289788 4888 generic.go:334] "Generic (PLEG): container finished" podID="e726333f-6370-454e-8cf8-79a3b8c72227" containerID="cbce5ece5719b7d88f8d6e16bacce79fd3e307a1604cc4623c49f68e1f724ab5" exitCode=0 Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.316220 4888 generic.go:334] "Generic (PLEG): container finished" podID="277e10a8-1c23-4683-9be9-0039e28cc55f" containerID="56eb51323f079c24a64eb9bf89416aa303002a66fb51a45a153b00120fa41b3b" exitCode=0 Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.316997 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"277e10a8-1c23-4683-9be9-0039e28cc55f","Type":"ContainerDied","Data":"56eb51323f079c24a64eb9bf89416aa303002a66fb51a45a153b00120fa41b3b"} Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.642055 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.754333 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access\") pod \"e726333f-6370-454e-8cf8-79a3b8c72227\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.754572 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir\") pod \"e726333f-6370-454e-8cf8-79a3b8c72227\" (UID: \"e726333f-6370-454e-8cf8-79a3b8c72227\") " Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.755400 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e726333f-6370-454e-8cf8-79a3b8c72227" (UID: "e726333f-6370-454e-8cf8-79a3b8c72227"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.798165 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e726333f-6370-454e-8cf8-79a3b8c72227" (UID: "e726333f-6370-454e-8cf8-79a3b8c72227"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.856492 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e726333f-6370-454e-8cf8-79a3b8c72227-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:32 crc kubenswrapper[4888]: I1124 00:27:32.856535 4888 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e726333f-6370-454e-8cf8-79a3b8c72227-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:33 crc kubenswrapper[4888]: I1124 00:27:33.380710 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 00:27:33 crc kubenswrapper[4888]: I1124 00:27:33.382892 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e726333f-6370-454e-8cf8-79a3b8c72227","Type":"ContainerDied","Data":"847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181"} Nov 24 00:27:33 crc kubenswrapper[4888]: I1124 00:27:33.382950 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="847e939073ed83471a6bcd0d875d1729cb2c9d03d0319a2f044f4ac3ae016181" Nov 24 00:27:34 crc kubenswrapper[4888]: I1124 00:27:34.839207 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-b52cm" Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.008888 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.014511 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.023579 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.023641 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.023858 4888 patch_prober.go:28] interesting pod/downloads-7954f5f757-5f942 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Nov 24 00:27:39 crc kubenswrapper[4888]: I1124 00:27:39.024287 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5f942" podUID="dec11a6c-0eb1-428a-b624-3558ffc42e1b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.28:8080/\": dial tcp 10.217.0.28:8080: connect: connection refused" Nov 24 00:27:40 crc kubenswrapper[4888]: I1124 00:27:40.904748 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:40 crc kubenswrapper[4888]: I1124 00:27:40.929501 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee556f32-31ee-45ef-bf08-26bf19390c96-metrics-certs\") pod \"network-metrics-daemon-7j4n7\" (UID: \"ee556f32-31ee-45ef-bf08-26bf19390c96\") " pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.140133 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.210306 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7j4n7" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.322193 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir\") pod \"277e10a8-1c23-4683-9be9-0039e28cc55f\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.322361 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access\") pod \"277e10a8-1c23-4683-9be9-0039e28cc55f\" (UID: \"277e10a8-1c23-4683-9be9-0039e28cc55f\") " Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.323608 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "277e10a8-1c23-4683-9be9-0039e28cc55f" (UID: "277e10a8-1c23-4683-9be9-0039e28cc55f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.327024 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "277e10a8-1c23-4683-9be9-0039e28cc55f" (UID: "277e10a8-1c23-4683-9be9-0039e28cc55f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.425805 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/277e10a8-1c23-4683-9be9-0039e28cc55f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.425867 4888 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/277e10a8-1c23-4683-9be9-0039e28cc55f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.449320 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"277e10a8-1c23-4683-9be9-0039e28cc55f","Type":"ContainerDied","Data":"b8f10a428d379d89290920deb89f705e7cd7127a7d7366a246dbcb1caddd79e0"} Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.449355 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 00:27:41 crc kubenswrapper[4888]: I1124 00:27:41.449366 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8f10a428d379d89290920deb89f705e7cd7127a7d7366a246dbcb1caddd79e0" Nov 24 00:27:44 crc kubenswrapper[4888]: I1124 00:27:44.974119 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7j4n7"] Nov 24 00:27:45 crc kubenswrapper[4888]: I1124 00:27:45.476923 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" event={"ID":"ee556f32-31ee-45ef-bf08-26bf19390c96","Type":"ContainerStarted","Data":"6571adb8596d16bd876a1a2bc84960fcaff624e45d418606f61c3a1b445d9187"} Nov 24 00:27:46 crc kubenswrapper[4888]: I1124 00:27:46.485083 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" event={"ID":"ee556f32-31ee-45ef-bf08-26bf19390c96","Type":"ContainerStarted","Data":"c5556eab9f47d008af4e9f29f3655168d7ee30165aef68bfeefa6119c6f9d618"} Nov 24 00:27:47 crc kubenswrapper[4888]: I1124 00:27:47.736243 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:27:49 crc kubenswrapper[4888]: I1124 00:27:49.035358 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5f942" Nov 24 00:27:53 crc kubenswrapper[4888]: I1124 00:27:53.530241 4888 generic.go:334] "Generic (PLEG): container finished" podID="8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" containerID="c92f8035bd03b463a08c6d01e451cf458d9cd4d65924dfea28ee5e717052e485" exitCode=0 Nov 24 00:27:53 crc kubenswrapper[4888]: I1124 00:27:53.530328 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29399040-nzntm" event={"ID":"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b","Type":"ContainerDied","Data":"c92f8035bd03b463a08c6d01e451cf458d9cd4d65924dfea28ee5e717052e485"} Nov 24 00:27:53 crc kubenswrapper[4888]: I1124 00:27:53.591522 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:27:53 crc kubenswrapper[4888]: I1124 00:27:53.591609 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.620032 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.758956 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca\") pod \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.759010 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpwgv\" (UniqueName: \"kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv\") pod \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\" (UID: \"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b\") " Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.759932 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca" (OuterVolumeSpecName: "serviceca") pod "8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" (UID: "8c6028fa-ae0e-47b7-bdf6-6f2367b6950b"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.766859 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv" (OuterVolumeSpecName: "kube-api-access-hpwgv") pod "8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" (UID: "8c6028fa-ae0e-47b7-bdf6-6f2367b6950b"). InnerVolumeSpecName "kube-api-access-hpwgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.861229 4888 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:56 crc kubenswrapper[4888]: I1124 00:27:56.862280 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpwgv\" (UniqueName: \"kubernetes.io/projected/8c6028fa-ae0e-47b7-bdf6-6f2367b6950b-kube-api-access-hpwgv\") on node \"crc\" DevicePath \"\"" Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.570338 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7j4n7" event={"ID":"ee556f32-31ee-45ef-bf08-26bf19390c96","Type":"ContainerStarted","Data":"6829a2bf3b0d610a27ca20149a3dc77a82e9eaf9e8ebbabb113655022c3d9d80"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.573607 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29399040-nzntm" Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.573639 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29399040-nzntm" event={"ID":"8c6028fa-ae0e-47b7-bdf6-6f2367b6950b","Type":"ContainerDied","Data":"1702902d58e3a5667e66c2ed60c75eb4e46ac522ac67ed8978bb6d811c17e054"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.573697 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1702902d58e3a5667e66c2ed60c75eb4e46ac522ac67ed8978bb6d811c17e054" Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.575635 4888 generic.go:334] "Generic (PLEG): container finished" podID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerID="85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.575688 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerDied","Data":"85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.583438 4888 generic.go:334] "Generic (PLEG): container finished" podID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerID="caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.583506 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerDied","Data":"caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.594557 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7j4n7" podStartSLOduration=159.59394706 podStartE2EDuration="2m39.59394706s" podCreationTimestamp="2025-11-24 00:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:57.584191894 +0000 UTC m=+180.166875958" watchObservedRunningTime="2025-11-24 00:27:57.59394706 +0000 UTC m=+180.176631104" Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.600162 4888 generic.go:334] "Generic (PLEG): container finished" podID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerID="4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.600255 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerDied","Data":"4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.603227 4888 generic.go:334] "Generic (PLEG): container finished" podID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerID="a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.603315 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerDied","Data":"a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.618170 4888 generic.go:334] "Generic (PLEG): container finished" podID="60936af3-7a00-49b5-8293-626b37cb41e5" containerID="9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.618274 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerDied","Data":"9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.627855 4888 generic.go:334] "Generic (PLEG): container finished" podID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerID="8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98" exitCode=0 Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.628047 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerDied","Data":"8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.638000 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerStarted","Data":"fe3f890c971a97782a406f6162bfaf3ff8ed6db4f5adfee09a059a8aac4605de"} Nov 24 00:27:57 crc kubenswrapper[4888]: I1124 00:27:57.643646 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerStarted","Data":"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006"} Nov 24 00:27:58 crc kubenswrapper[4888]: I1124 00:27:58.654168 4888 generic.go:334] "Generic (PLEG): container finished" podID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerID="fe3f890c971a97782a406f6162bfaf3ff8ed6db4f5adfee09a059a8aac4605de" exitCode=0 Nov 24 00:27:58 crc kubenswrapper[4888]: I1124 00:27:58.654248 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerDied","Data":"fe3f890c971a97782a406f6162bfaf3ff8ed6db4f5adfee09a059a8aac4605de"} Nov 24 00:27:58 crc kubenswrapper[4888]: I1124 00:27:58.657045 4888 generic.go:334] "Generic (PLEG): container finished" podID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerID="49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006" exitCode=0 Nov 24 00:27:58 crc kubenswrapper[4888]: I1124 00:27:58.657183 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerDied","Data":"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006"} Nov 24 00:28:00 crc kubenswrapper[4888]: I1124 00:28:00.019766 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7jtk" Nov 24 00:28:00 crc kubenswrapper[4888]: I1124 00:28:00.671584 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerStarted","Data":"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4"} Nov 24 00:28:00 crc kubenswrapper[4888]: I1124 00:28:00.676896 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerStarted","Data":"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953"} Nov 24 00:28:00 crc kubenswrapper[4888]: I1124 00:28:00.707941 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ngz8m" podStartSLOduration=3.646317408 podStartE2EDuration="35.707914538s" podCreationTimestamp="2025-11-24 00:27:25 +0000 UTC" firstStartedPulling="2025-11-24 00:27:28.014979052 +0000 UTC m=+150.597663096" lastFinishedPulling="2025-11-24 00:28:00.076576182 +0000 UTC m=+182.659260226" observedRunningTime="2025-11-24 00:28:00.70334027 +0000 UTC m=+183.286024324" watchObservedRunningTime="2025-11-24 00:28:00.707914538 +0000 UTC m=+183.290598592" Nov 24 00:28:01 crc kubenswrapper[4888]: I1124 00:28:01.702835 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jknzx" podStartSLOduration=3.79594421 podStartE2EDuration="33.702797331s" podCreationTimestamp="2025-11-24 00:27:28 +0000 UTC" firstStartedPulling="2025-11-24 00:27:30.208431657 +0000 UTC m=+152.791115701" lastFinishedPulling="2025-11-24 00:28:00.115284778 +0000 UTC m=+182.697968822" observedRunningTime="2025-11-24 00:28:01.700407793 +0000 UTC m=+184.283091847" watchObservedRunningTime="2025-11-24 00:28:01.702797331 +0000 UTC m=+184.285481385" Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.691646 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerStarted","Data":"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057"} Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.694513 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerStarted","Data":"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f"} Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.698297 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerStarted","Data":"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0"} Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.701067 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerStarted","Data":"65824cb5a422e66a1e173280ee426260e050525980d232c7e722643350dd70ff"} Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.704016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerStarted","Data":"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34"} Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.718085 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sg6wm" podStartSLOduration=3.699541785 podStartE2EDuration="37.718070466s" podCreationTimestamp="2025-11-24 00:27:25 +0000 UTC" firstStartedPulling="2025-11-24 00:27:27.996314266 +0000 UTC m=+150.578998310" lastFinishedPulling="2025-11-24 00:28:02.014842947 +0000 UTC m=+184.597526991" observedRunningTime="2025-11-24 00:28:02.716562777 +0000 UTC m=+185.299246821" watchObservedRunningTime="2025-11-24 00:28:02.718070466 +0000 UTC m=+185.300754510" Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.747716 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hvhgl" podStartSLOduration=2.9188280989999997 podStartE2EDuration="33.747700478s" podCreationTimestamp="2025-11-24 00:27:29 +0000 UTC" firstStartedPulling="2025-11-24 00:27:31.288804265 +0000 UTC m=+153.871488309" lastFinishedPulling="2025-11-24 00:28:02.117676644 +0000 UTC m=+184.700360688" observedRunningTime="2025-11-24 00:28:02.747275164 +0000 UTC m=+185.329959208" watchObservedRunningTime="2025-11-24 00:28:02.747700478 +0000 UTC m=+185.330384512" Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.771734 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9hrl6" podStartSLOduration=2.911981429 podStartE2EDuration="36.771714087s" podCreationTimestamp="2025-11-24 00:27:26 +0000 UTC" firstStartedPulling="2025-11-24 00:27:28.042177204 +0000 UTC m=+150.624861248" lastFinishedPulling="2025-11-24 00:28:01.901909872 +0000 UTC m=+184.484593906" observedRunningTime="2025-11-24 00:28:02.7674843 +0000 UTC m=+185.350168364" watchObservedRunningTime="2025-11-24 00:28:02.771714087 +0000 UTC m=+185.354398131" Nov 24 00:28:02 crc kubenswrapper[4888]: I1124 00:28:02.793153 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7fx6r" podStartSLOduration=3.054178672 podStartE2EDuration="35.793126842s" podCreationTimestamp="2025-11-24 00:27:27 +0000 UTC" firstStartedPulling="2025-11-24 00:27:29.14224084 +0000 UTC m=+151.724924884" lastFinishedPulling="2025-11-24 00:28:01.88118899 +0000 UTC m=+184.463873054" observedRunningTime="2025-11-24 00:28:02.790763835 +0000 UTC m=+185.373447879" watchObservedRunningTime="2025-11-24 00:28:02.793126842 +0000 UTC m=+185.375810886" Nov 24 00:28:03 crc kubenswrapper[4888]: I1124 00:28:03.713951 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerStarted","Data":"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779"} Nov 24 00:28:03 crc kubenswrapper[4888]: I1124 00:28:03.735544 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fgpsq" podStartSLOduration=3.820748446 podStartE2EDuration="35.735500771s" podCreationTimestamp="2025-11-24 00:27:28 +0000 UTC" firstStartedPulling="2025-11-24 00:27:31.287366258 +0000 UTC m=+153.870050302" lastFinishedPulling="2025-11-24 00:28:03.202118583 +0000 UTC m=+185.784802627" observedRunningTime="2025-11-24 00:28:03.734749337 +0000 UTC m=+186.317433381" watchObservedRunningTime="2025-11-24 00:28:03.735500771 +0000 UTC m=+186.318184805" Nov 24 00:28:03 crc kubenswrapper[4888]: I1124 00:28:03.760334 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mkctm" podStartSLOduration=4.7329896399999996 podStartE2EDuration="38.760310616s" podCreationTimestamp="2025-11-24 00:27:25 +0000 UTC" firstStartedPulling="2025-11-24 00:27:28.056962814 +0000 UTC m=+150.639646858" lastFinishedPulling="2025-11-24 00:28:02.08428377 +0000 UTC m=+184.666967834" observedRunningTime="2025-11-24 00:28:03.758149246 +0000 UTC m=+186.340833300" watchObservedRunningTime="2025-11-24 00:28:03.760310616 +0000 UTC m=+186.342994660" Nov 24 00:28:05 crc kubenswrapper[4888]: I1124 00:28:05.478367 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 00:28:05 crc kubenswrapper[4888]: I1124 00:28:05.899051 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:05 crc kubenswrapper[4888]: I1124 00:28:05.899161 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.152005 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.152068 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.341571 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.341648 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.362633 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.367869 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.397885 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.562575 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.562637 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.615576 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:06 crc kubenswrapper[4888]: I1124 00:28:06.785514 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.025701 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.081962 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.082020 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.202323 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.478030 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.478094 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.533855 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.807463 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:08 crc kubenswrapper[4888]: I1124 00:28:08.811038 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.024092 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.024401 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.498077 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.498154 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.547131 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:09 crc kubenswrapper[4888]: I1124 00:28:09.807910 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:10 crc kubenswrapper[4888]: I1124 00:28:10.084646 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fgpsq" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="registry-server" probeResult="failure" output=< Nov 24 00:28:10 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:28:10 crc kubenswrapper[4888]: > Nov 24 00:28:11 crc kubenswrapper[4888]: I1124 00:28:11.913180 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:28:11 crc kubenswrapper[4888]: I1124 00:28:11.913521 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jknzx" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="registry-server" containerID="cri-o://4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4" gracePeriod=2 Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.268915 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.439371 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content\") pod \"60936af3-7a00-49b5-8293-626b37cb41e5\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.439529 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwqks\" (UniqueName: \"kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks\") pod \"60936af3-7a00-49b5-8293-626b37cb41e5\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.439579 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities\") pod \"60936af3-7a00-49b5-8293-626b37cb41e5\" (UID: \"60936af3-7a00-49b5-8293-626b37cb41e5\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.440490 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities" (OuterVolumeSpecName: "utilities") pod "60936af3-7a00-49b5-8293-626b37cb41e5" (UID: "60936af3-7a00-49b5-8293-626b37cb41e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.449243 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks" (OuterVolumeSpecName: "kube-api-access-fwqks") pod "60936af3-7a00-49b5-8293-626b37cb41e5" (UID: "60936af3-7a00-49b5-8293-626b37cb41e5"). InnerVolumeSpecName "kube-api-access-fwqks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.456622 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60936af3-7a00-49b5-8293-626b37cb41e5" (UID: "60936af3-7a00-49b5-8293-626b37cb41e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.511774 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.512109 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hvhgl" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="registry-server" containerID="cri-o://65824cb5a422e66a1e173280ee426260e050525980d232c7e722643350dd70ff" gracePeriod=2 Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.541706 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwqks\" (UniqueName: \"kubernetes.io/projected/60936af3-7a00-49b5-8293-626b37cb41e5-kube-api-access-fwqks\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.541747 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.541759 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60936af3-7a00-49b5-8293-626b37cb41e5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.786598 4888 generic.go:334] "Generic (PLEG): container finished" podID="60936af3-7a00-49b5-8293-626b37cb41e5" containerID="4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4" exitCode=0 Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.786663 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerDied","Data":"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4"} Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.786698 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jknzx" event={"ID":"60936af3-7a00-49b5-8293-626b37cb41e5","Type":"ContainerDied","Data":"86cfe8b83975ceeed94b8c24a6da0749ad38946779cc66360e6608a1bb323f9d"} Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.786719 4888 scope.go:117] "RemoveContainer" containerID="4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.786852 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jknzx" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.792019 4888 generic.go:334] "Generic (PLEG): container finished" podID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerID="65824cb5a422e66a1e173280ee426260e050525980d232c7e722643350dd70ff" exitCode=0 Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.792047 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerDied","Data":"65824cb5a422e66a1e173280ee426260e050525980d232c7e722643350dd70ff"} Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.824388 4888 scope.go:117] "RemoveContainer" containerID="9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.827483 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.828054 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.831500 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jknzx"] Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.845962 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities\") pod \"c54474df-3d22-4c12-a402-47fb8d8ee014\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.848209 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities" (OuterVolumeSpecName: "utilities") pod "c54474df-3d22-4c12-a402-47fb8d8ee014" (UID: "c54474df-3d22-4c12-a402-47fb8d8ee014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.853055 4888 scope.go:117] "RemoveContainer" containerID="53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.870771 4888 scope.go:117] "RemoveContainer" containerID="4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4" Nov 24 00:28:12 crc kubenswrapper[4888]: E1124 00:28:12.871325 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4\": container with ID starting with 4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4 not found: ID does not exist" containerID="4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.871362 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4"} err="failed to get container status \"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4\": rpc error: code = NotFound desc = could not find container \"4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4\": container with ID starting with 4de29af12e6c50f42ff330532431556c825c5556f358b6c10be02c55a9a28dd4 not found: ID does not exist" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.871415 4888 scope.go:117] "RemoveContainer" containerID="9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455" Nov 24 00:28:12 crc kubenswrapper[4888]: E1124 00:28:12.871853 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455\": container with ID starting with 9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455 not found: ID does not exist" containerID="9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.871909 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455"} err="failed to get container status \"9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455\": rpc error: code = NotFound desc = could not find container \"9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455\": container with ID starting with 9b82c7e1f40372a12a0956c38817f7ef6d7248c8ce9390b630b7af8138cb4455 not found: ID does not exist" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.871949 4888 scope.go:117] "RemoveContainer" containerID="53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14" Nov 24 00:28:12 crc kubenswrapper[4888]: E1124 00:28:12.872360 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14\": container with ID starting with 53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14 not found: ID does not exist" containerID="53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.872399 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14"} err="failed to get container status \"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14\": rpc error: code = NotFound desc = could not find container \"53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14\": container with ID starting with 53d6a36f48e20b2c7aa078be6367a591c09e4b3e1c378b92939573c9ea198c14 not found: ID does not exist" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.947226 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwwxj\" (UniqueName: \"kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj\") pod \"c54474df-3d22-4c12-a402-47fb8d8ee014\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.947475 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content\") pod \"c54474df-3d22-4c12-a402-47fb8d8ee014\" (UID: \"c54474df-3d22-4c12-a402-47fb8d8ee014\") " Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.947750 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:12 crc kubenswrapper[4888]: I1124 00:28:12.951717 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj" (OuterVolumeSpecName: "kube-api-access-dwwxj") pod "c54474df-3d22-4c12-a402-47fb8d8ee014" (UID: "c54474df-3d22-4c12-a402-47fb8d8ee014"). InnerVolumeSpecName "kube-api-access-dwwxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.049315 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwwxj\" (UniqueName: \"kubernetes.io/projected/c54474df-3d22-4c12-a402-47fb8d8ee014-kube-api-access-dwwxj\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.052102 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c54474df-3d22-4c12-a402-47fb8d8ee014" (UID: "c54474df-3d22-4c12-a402-47fb8d8ee014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.151630 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c54474df-3d22-4c12-a402-47fb8d8ee014-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.801510 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvhgl" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.801535 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvhgl" event={"ID":"c54474df-3d22-4c12-a402-47fb8d8ee014","Type":"ContainerDied","Data":"4c3e6e991f2a2d17dd837f06432bb35e16ff1de80e51aaba11f42d3f7572de44"} Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.801958 4888 scope.go:117] "RemoveContainer" containerID="65824cb5a422e66a1e173280ee426260e050525980d232c7e722643350dd70ff" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.834216 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.839944 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hvhgl"] Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.843669 4888 scope.go:117] "RemoveContainer" containerID="fe3f890c971a97782a406f6162bfaf3ff8ed6db4f5adfee09a059a8aac4605de" Nov 24 00:28:13 crc kubenswrapper[4888]: I1124 00:28:13.873273 4888 scope.go:117] "RemoveContainer" containerID="f6d0ef01e87f4837f43e15f54e5a8b5bf29a45f516dbdb818afb6a9a73179189" Nov 24 00:28:14 crc kubenswrapper[4888]: I1124 00:28:14.252984 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" path="/var/lib/kubelet/pods/60936af3-7a00-49b5-8293-626b37cb41e5/volumes" Nov 24 00:28:14 crc kubenswrapper[4888]: I1124 00:28:14.254261 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" path="/var/lib/kubelet/pods/c54474df-3d22-4c12-a402-47fb8d8ee014/volumes" Nov 24 00:28:16 crc kubenswrapper[4888]: I1124 00:28:16.190664 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:16 crc kubenswrapper[4888]: I1124 00:28:16.397963 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:16 crc kubenswrapper[4888]: I1124 00:28:16.609848 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:18 crc kubenswrapper[4888]: I1124 00:28:18.915038 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:28:18 crc kubenswrapper[4888]: I1124 00:28:18.915357 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sg6wm" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="registry-server" containerID="cri-o://3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057" gracePeriod=2 Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.098906 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.155188 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.388918 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.549498 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content\") pod \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.549630 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwtq6\" (UniqueName: \"kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6\") pod \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.549674 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities\") pod \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\" (UID: \"fd19d281-87bb-463f-b58b-7a5f80a8eb3d\") " Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.550654 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities" (OuterVolumeSpecName: "utilities") pod "fd19d281-87bb-463f-b58b-7a5f80a8eb3d" (UID: "fd19d281-87bb-463f-b58b-7a5f80a8eb3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.558578 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6" (OuterVolumeSpecName: "kube-api-access-hwtq6") pod "fd19d281-87bb-463f-b58b-7a5f80a8eb3d" (UID: "fd19d281-87bb-463f-b58b-7a5f80a8eb3d"). InnerVolumeSpecName "kube-api-access-hwtq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.601671 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd19d281-87bb-463f-b58b-7a5f80a8eb3d" (UID: "fd19d281-87bb-463f-b58b-7a5f80a8eb3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.663439 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.663508 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwtq6\" (UniqueName: \"kubernetes.io/projected/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-kube-api-access-hwtq6\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.663554 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd19d281-87bb-463f-b58b-7a5f80a8eb3d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.846995 4888 generic.go:334] "Generic (PLEG): container finished" podID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerID="3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057" exitCode=0 Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.847101 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sg6wm" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.847141 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerDied","Data":"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057"} Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.847243 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sg6wm" event={"ID":"fd19d281-87bb-463f-b58b-7a5f80a8eb3d","Type":"ContainerDied","Data":"4e4454eb035152a7ba54cfe14508275452580a8b0a6592242c4780203cbca785"} Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.847287 4888 scope.go:117] "RemoveContainer" containerID="3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.874044 4888 scope.go:117] "RemoveContainer" containerID="8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.892771 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.897621 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sg6wm"] Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.906115 4888 scope.go:117] "RemoveContainer" containerID="96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.931486 4888 scope.go:117] "RemoveContainer" containerID="3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057" Nov 24 00:28:19 crc kubenswrapper[4888]: E1124 00:28:19.932163 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057\": container with ID starting with 3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057 not found: ID does not exist" containerID="3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.932198 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057"} err="failed to get container status \"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057\": rpc error: code = NotFound desc = could not find container \"3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057\": container with ID starting with 3dd935c196e61eac82a099a720cab31a2565c3f9d91a011b1653333ad7cc3057 not found: ID does not exist" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.932225 4888 scope.go:117] "RemoveContainer" containerID="8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98" Nov 24 00:28:19 crc kubenswrapper[4888]: E1124 00:28:19.932527 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98\": container with ID starting with 8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98 not found: ID does not exist" containerID="8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.932549 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98"} err="failed to get container status \"8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98\": rpc error: code = NotFound desc = could not find container \"8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98\": container with ID starting with 8c23ba4619bc8d356a5e2f93af6543559f4155a25a2e3f1aa0dbcacece0bfd98 not found: ID does not exist" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.932563 4888 scope.go:117] "RemoveContainer" containerID="96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8" Nov 24 00:28:19 crc kubenswrapper[4888]: E1124 00:28:19.932873 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8\": container with ID starting with 96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8 not found: ID does not exist" containerID="96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8" Nov 24 00:28:19 crc kubenswrapper[4888]: I1124 00:28:19.932898 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8"} err="failed to get container status \"96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8\": rpc error: code = NotFound desc = could not find container \"96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8\": container with ID starting with 96b9d5001838626f851eb3109145c4d8f355f1f3aba16ede24c93e0d32489ab8 not found: ID does not exist" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.255557 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" path="/var/lib/kubelet/pods/fd19d281-87bb-463f-b58b-7a5f80a8eb3d/volumes" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.320390 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.320802 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9hrl6" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="registry-server" containerID="cri-o://b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0" gracePeriod=2 Nov 24 00:28:20 crc kubenswrapper[4888]: E1124 00:28:20.443138 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddece2a22_2482_44ff_b1e6_237ce30f3cb1.slice/crio-b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.718034 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.855245 4888 generic.go:334] "Generic (PLEG): container finished" podID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerID="b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0" exitCode=0 Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.855317 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerDied","Data":"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0"} Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.855325 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hrl6" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.855368 4888 scope.go:117] "RemoveContainer" containerID="b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.855355 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hrl6" event={"ID":"dece2a22-2482-44ff-b1e6-237ce30f3cb1","Type":"ContainerDied","Data":"ed9d76efc3c3c7b33f364019e12ebadd49c6412914c33961d77a6816beaa0d45"} Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.872729 4888 scope.go:117] "RemoveContainer" containerID="4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.881786 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities\") pod \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.881923 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content\") pod \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.882110 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkmzq\" (UniqueName: \"kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq\") pod \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\" (UID: \"dece2a22-2482-44ff-b1e6-237ce30f3cb1\") " Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.882429 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities" (OuterVolumeSpecName: "utilities") pod "dece2a22-2482-44ff-b1e6-237ce30f3cb1" (UID: "dece2a22-2482-44ff-b1e6-237ce30f3cb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.882586 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.888555 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq" (OuterVolumeSpecName: "kube-api-access-dkmzq") pod "dece2a22-2482-44ff-b1e6-237ce30f3cb1" (UID: "dece2a22-2482-44ff-b1e6-237ce30f3cb1"). InnerVolumeSpecName "kube-api-access-dkmzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.905224 4888 scope.go:117] "RemoveContainer" containerID="646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.935708 4888 scope.go:117] "RemoveContainer" containerID="b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0" Nov 24 00:28:20 crc kubenswrapper[4888]: E1124 00:28:20.937717 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0\": container with ID starting with b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0 not found: ID does not exist" containerID="b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.937786 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0"} err="failed to get container status \"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0\": rpc error: code = NotFound desc = could not find container \"b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0\": container with ID starting with b9875b2afd2e41d7c62d46c43bf050b7b310909295f083f984744b289e588eb0 not found: ID does not exist" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.937861 4888 scope.go:117] "RemoveContainer" containerID="4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63" Nov 24 00:28:20 crc kubenswrapper[4888]: E1124 00:28:20.938938 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63\": container with ID starting with 4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63 not found: ID does not exist" containerID="4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.938986 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63"} err="failed to get container status \"4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63\": rpc error: code = NotFound desc = could not find container \"4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63\": container with ID starting with 4962abb0e43e9124a84c792a2002f87d5e66a15df07a9e669431723cb333ca63 not found: ID does not exist" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.939011 4888 scope.go:117] "RemoveContainer" containerID="646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d" Nov 24 00:28:20 crc kubenswrapper[4888]: E1124 00:28:20.939504 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d\": container with ID starting with 646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d not found: ID does not exist" containerID="646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.939569 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d"} err="failed to get container status \"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d\": rpc error: code = NotFound desc = could not find container \"646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d\": container with ID starting with 646abcf1ea97d59d7c7ae67b002b7c33d297a08b0dffca29356c3e4c29093c9d not found: ID does not exist" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.943602 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dece2a22-2482-44ff-b1e6-237ce30f3cb1" (UID: "dece2a22-2482-44ff-b1e6-237ce30f3cb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.984223 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dece2a22-2482-44ff-b1e6-237ce30f3cb1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:20 crc kubenswrapper[4888]: I1124 00:28:20.984492 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkmzq\" (UniqueName: \"kubernetes.io/projected/dece2a22-2482-44ff-b1e6-237ce30f3cb1-kube-api-access-dkmzq\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:21 crc kubenswrapper[4888]: I1124 00:28:21.189307 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:28:21 crc kubenswrapper[4888]: I1124 00:28:21.207280 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9hrl6"] Nov 24 00:28:22 crc kubenswrapper[4888]: I1124 00:28:22.257722 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" path="/var/lib/kubelet/pods/dece2a22-2482-44ff-b1e6-237ce30f3cb1/volumes" Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.592347 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.593064 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.593147 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.594386 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.594531 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7" gracePeriod=600 Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.884022 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7" exitCode=0 Nov 24 00:28:23 crc kubenswrapper[4888]: I1124 00:28:23.884165 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7"} Nov 24 00:28:24 crc kubenswrapper[4888]: I1124 00:28:24.892694 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4"} Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.062225 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerName="oauth-openshift" containerID="cri-o://c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b" gracePeriod=15 Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.534258 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598107 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598167 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598188 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598217 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598234 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598608 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598720 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598750 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598775 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598826 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5qnx\" (UniqueName: \"kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598844 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598864 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.598890 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template\") pod \"b847db54-4cfd-4b84-b3a8-a8aad4304919\" (UID: \"b847db54-4cfd-4b84-b3a8-a8aad4304919\") " Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.603079 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.603607 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.604412 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.605227 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.605493 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.609281 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.609771 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.612505 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.612552 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.613247 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.613240 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx" (OuterVolumeSpecName: "kube-api-access-f5qnx") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "kube-api-access-f5qnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.613589 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.615195 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.621167 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b847db54-4cfd-4b84-b3a8-a8aad4304919" (UID: "b847db54-4cfd-4b84-b3a8-a8aad4304919"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699802 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699909 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5qnx\" (UniqueName: \"kubernetes.io/projected/b847db54-4cfd-4b84-b3a8-a8aad4304919-kube-api-access-f5qnx\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699925 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699936 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699945 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699954 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699964 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699977 4888 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699986 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.699995 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.700004 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.700013 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.700022 4888 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b847db54-4cfd-4b84-b3a8-a8aad4304919-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.700031 4888 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b847db54-4cfd-4b84-b3a8-a8aad4304919-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.949708 4888 generic.go:334] "Generic (PLEG): container finished" podID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerID="c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b" exitCode=0 Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.949777 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" event={"ID":"b847db54-4cfd-4b84-b3a8-a8aad4304919","Type":"ContainerDied","Data":"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b"} Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.949782 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.949840 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-st95l" event={"ID":"b847db54-4cfd-4b84-b3a8-a8aad4304919","Type":"ContainerDied","Data":"51b68b26ea86aad561dd9cc75c99a4f20733e8363173e6e7e24f84da05b82d86"} Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.949862 4888 scope.go:117] "RemoveContainer" containerID="c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.972620 4888 scope.go:117] "RemoveContainer" containerID="c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b" Nov 24 00:28:33 crc kubenswrapper[4888]: E1124 00:28:33.973687 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b\": container with ID starting with c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b not found: ID does not exist" containerID="c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.973729 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b"} err="failed to get container status \"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b\": rpc error: code = NotFound desc = could not find container \"c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b\": container with ID starting with c99aea0d4793a2153844a129eda019dfbd592471d264a01a0dcdbeeaa2f3e37b not found: ID does not exist" Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.985187 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:28:33 crc kubenswrapper[4888]: I1124 00:28:33.996393 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-st95l"] Nov 24 00:28:34 crc kubenswrapper[4888]: I1124 00:28:34.259108 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" path="/var/lib/kubelet/pods/b847db54-4cfd-4b84-b3a8-a8aad4304919/volumes" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.361771 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79798c6d69-skhx8"] Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362412 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e726333f-6370-454e-8cf8-79a3b8c72227" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362431 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e726333f-6370-454e-8cf8-79a3b8c72227" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362442 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362449 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362458 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerName="oauth-openshift" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362464 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerName="oauth-openshift" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362482 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362490 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362501 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362509 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362522 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" containerName="image-pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362530 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" containerName="image-pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362543 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362550 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362560 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362567 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362577 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362585 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362596 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277e10a8-1c23-4683-9be9-0039e28cc55f" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362605 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="277e10a8-1c23-4683-9be9-0039e28cc55f" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362617 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362627 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362640 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362647 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="extract-utilities" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362658 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362666 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362676 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362684 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="extract-content" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362696 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153d6369-b587-4da4-863c-bc8a85edde5e" containerName="collect-profiles" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362703 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="153d6369-b587-4da4-863c-bc8a85edde5e" containerName="collect-profiles" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362713 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362720 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: E1124 00:28:38.362729 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362736 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362882 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="277e10a8-1c23-4683-9be9-0039e28cc55f" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362895 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c6028fa-ae0e-47b7-bdf6-6f2367b6950b" containerName="image-pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362904 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54474df-3d22-4c12-a402-47fb8d8ee014" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362912 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b847db54-4cfd-4b84-b3a8-a8aad4304919" containerName="oauth-openshift" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362924 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="153d6369-b587-4da4-863c-bc8a85edde5e" containerName="collect-profiles" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362940 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd19d281-87bb-463f-b58b-7a5f80a8eb3d" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362951 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e726333f-6370-454e-8cf8-79a3b8c72227" containerName="pruner" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362960 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="60936af3-7a00-49b5-8293-626b37cb41e5" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.362969 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dece2a22-2482-44ff-b1e6-237ce30f3cb1" containerName="registry-server" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.363510 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.368407 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.369194 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.371288 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.371618 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.382269 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.382953 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.383177 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.385096 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.385257 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.386040 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.386578 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.390308 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392140 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-service-ca\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392286 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73764f8a-00d8-472b-bdd3-5025fee55893-audit-dir\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392406 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392631 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-router-certs\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392694 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392861 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.392950 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-audit-policies\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393028 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-error\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393317 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393484 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-login\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393633 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-session\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.393947 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw7gh\" (UniqueName: \"kubernetes.io/projected/73764f8a-00d8-472b-bdd3-5025fee55893-kube-api-access-kw7gh\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.394282 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.413368 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.416719 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79798c6d69-skhx8"] Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.433749 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.435206 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497713 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497774 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-login\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497827 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-session\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497857 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw7gh\" (UniqueName: \"kubernetes.io/projected/73764f8a-00d8-472b-bdd3-5025fee55893-kube-api-access-kw7gh\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497887 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497929 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-service-ca\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.497965 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73764f8a-00d8-472b-bdd3-5025fee55893-audit-dir\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498030 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-router-certs\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498055 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498091 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498119 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-audit-policies\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498153 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-error\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498185 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.498918 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73764f8a-00d8-472b-bdd3-5025fee55893-audit-dir\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.499834 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.499990 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.500109 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-service-ca\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.500613 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73764f8a-00d8-472b-bdd3-5025fee55893-audit-policies\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.506400 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-session\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.507628 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-router-certs\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.507979 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.508257 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.508456 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.509214 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.510622 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-login\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.519175 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/73764f8a-00d8-472b-bdd3-5025fee55893-v4-0-config-user-template-error\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.525171 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw7gh\" (UniqueName: \"kubernetes.io/projected/73764f8a-00d8-472b-bdd3-5025fee55893-kube-api-access-kw7gh\") pod \"oauth-openshift-79798c6d69-skhx8\" (UID: \"73764f8a-00d8-472b-bdd3-5025fee55893\") " pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.718595 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:38 crc kubenswrapper[4888]: I1124 00:28:38.996450 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79798c6d69-skhx8"] Nov 24 00:28:39 crc kubenswrapper[4888]: I1124 00:28:39.997227 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" event={"ID":"73764f8a-00d8-472b-bdd3-5025fee55893","Type":"ContainerStarted","Data":"614ed5aee1da5e66f5d9f5495ed359013a84895527572b9e9ca4e9e0700b891f"} Nov 24 00:28:39 crc kubenswrapper[4888]: I1124 00:28:39.998391 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" event={"ID":"73764f8a-00d8-472b-bdd3-5025fee55893","Type":"ContainerStarted","Data":"ad5cc040bbcbf64a99abf7b9bea5d254d98dd095a63291eeb3a7d8a8a4f55d40"} Nov 24 00:28:39 crc kubenswrapper[4888]: I1124 00:28:39.998436 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:40 crc kubenswrapper[4888]: I1124 00:28:40.008570 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" Nov 24 00:28:40 crc kubenswrapper[4888]: I1124 00:28:40.029080 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79798c6d69-skhx8" podStartSLOduration=32.029045116 podStartE2EDuration="32.029045116s" podCreationTimestamp="2025-11-24 00:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:28:40.027492483 +0000 UTC m=+222.610176567" watchObservedRunningTime="2025-11-24 00:28:40.029045116 +0000 UTC m=+222.611729190" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.489241 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.490582 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ngz8m" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="registry-server" containerID="cri-o://faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953" gracePeriod=30 Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.517876 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.518420 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mkctm" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="registry-server" containerID="cri-o://8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34" gracePeriod=30 Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.530748 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.531041 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" containerID="cri-o://d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e" gracePeriod=30 Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.539157 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.539541 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7fx6r" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="registry-server" containerID="cri-o://c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f" gracePeriod=30 Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.546509 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.546876 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fgpsq" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="registry-server" containerID="cri-o://86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779" gracePeriod=30 Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.551187 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nmmd"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.552213 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.569410 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nmmd"] Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.621521 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.621586 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.621654 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crqpn\" (UniqueName: \"kubernetes.io/projected/ef22eb11-250d-44e2-a2bf-7906d7817b8b-kube-api-access-crqpn\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.723269 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crqpn\" (UniqueName: \"kubernetes.io/projected/ef22eb11-250d-44e2-a2bf-7906d7817b8b-kube-api-access-crqpn\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.723409 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.723432 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.725477 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.748485 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ef22eb11-250d-44e2-a2bf-7906d7817b8b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.753441 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crqpn\" (UniqueName: \"kubernetes.io/projected/ef22eb11-250d-44e2-a2bf-7906d7817b8b-kube-api-access-crqpn\") pod \"marketplace-operator-79b997595-4nmmd\" (UID: \"ef22eb11-250d-44e2-a2bf-7906d7817b8b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.960899 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:51 crc kubenswrapper[4888]: I1124 00:28:51.970287 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.028325 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content\") pod \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.028425 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4rwb\" (UniqueName: \"kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb\") pod \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.028472 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities\") pod \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\" (UID: \"8c084ed8-8f40-4bba-b2d8-07992d26fd35\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.032485 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.032700 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities" (OuterVolumeSpecName: "utilities") pod "8c084ed8-8f40-4bba-b2d8-07992d26fd35" (UID: "8c084ed8-8f40-4bba-b2d8-07992d26fd35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.045231 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb" (OuterVolumeSpecName: "kube-api-access-b4rwb") pod "8c084ed8-8f40-4bba-b2d8-07992d26fd35" (UID: "8c084ed8-8f40-4bba-b2d8-07992d26fd35"). InnerVolumeSpecName "kube-api-access-b4rwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.063546 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.101332 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.104166 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.119740 4888 generic.go:334] "Generic (PLEG): container finished" podID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerID="d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e" exitCode=0 Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.119820 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" event={"ID":"859fc02a-e8f2-4aec-99ce-d00f153ac581","Type":"ContainerDied","Data":"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.119855 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" event={"ID":"859fc02a-e8f2-4aec-99ce-d00f153ac581","Type":"ContainerDied","Data":"4e7976bfa5fe59cd85a2e6cbd41694eee02e46ebaf9212efd477f5c739000a75"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.119879 4888 scope.go:117] "RemoveContainer" containerID="d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.120005 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7vpcc" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.123731 4888 generic.go:334] "Generic (PLEG): container finished" podID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerID="faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953" exitCode=0 Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.123782 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngz8m" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.123847 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerDied","Data":"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.123880 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngz8m" event={"ID":"8c084ed8-8f40-4bba-b2d8-07992d26fd35","Type":"ContainerDied","Data":"bfc32be6325894c640a0441a2ba0c44f1e50262b7493137bb7be1150d86a9e1b"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.131863 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content\") pod \"58e54292-9a18-4074-aa87-a9d3a8372d49\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.132411 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics\") pod \"859fc02a-e8f2-4aec-99ce-d00f153ac581\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.132588 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nksrp\" (UniqueName: \"kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp\") pod \"58e54292-9a18-4074-aa87-a9d3a8372d49\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.132755 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities\") pod \"58e54292-9a18-4074-aa87-a9d3a8372d49\" (UID: \"58e54292-9a18-4074-aa87-a9d3a8372d49\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133165 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ns44\" (UniqueName: \"kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44\") pod \"859fc02a-e8f2-4aec-99ce-d00f153ac581\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133306 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities\") pod \"a7523f74-b2dc-4eaa-a12e-e244510a652b\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133741 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vch5n\" (UniqueName: \"kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n\") pod \"a7523f74-b2dc-4eaa-a12e-e244510a652b\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133845 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srsj7\" (UniqueName: \"kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7\") pod \"0694eb5e-cf26-46a4-b695-91706ed2fb13\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133911 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca\") pod \"859fc02a-e8f2-4aec-99ce-d00f153ac581\" (UID: \"859fc02a-e8f2-4aec-99ce-d00f153ac581\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.133947 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content\") pod \"a7523f74-b2dc-4eaa-a12e-e244510a652b\" (UID: \"a7523f74-b2dc-4eaa-a12e-e244510a652b\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.134068 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content\") pod \"0694eb5e-cf26-46a4-b695-91706ed2fb13\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.134121 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities\") pod \"0694eb5e-cf26-46a4-b695-91706ed2fb13\" (UID: \"0694eb5e-cf26-46a4-b695-91706ed2fb13\") " Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.135613 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.135642 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4rwb\" (UniqueName: \"kubernetes.io/projected/8c084ed8-8f40-4bba-b2d8-07992d26fd35-kube-api-access-b4rwb\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.136078 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities" (OuterVolumeSpecName: "utilities") pod "a7523f74-b2dc-4eaa-a12e-e244510a652b" (UID: "a7523f74-b2dc-4eaa-a12e-e244510a652b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.136398 4888 generic.go:334] "Generic (PLEG): container finished" podID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerID="c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f" exitCode=0 Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.136519 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerDied","Data":"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.136845 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fx6r" event={"ID":"a7523f74-b2dc-4eaa-a12e-e244510a652b","Type":"ContainerDied","Data":"7f18008515a37eef341febcd0e01bc031b1e5a2646e1cb6dcf677332ba168deb"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.136960 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fx6r" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.138509 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "859fc02a-e8f2-4aec-99ce-d00f153ac581" (UID: "859fc02a-e8f2-4aec-99ce-d00f153ac581"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.142368 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities" (OuterVolumeSpecName: "utilities") pod "58e54292-9a18-4074-aa87-a9d3a8372d49" (UID: "58e54292-9a18-4074-aa87-a9d3a8372d49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.144110 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities" (OuterVolumeSpecName: "utilities") pod "0694eb5e-cf26-46a4-b695-91706ed2fb13" (UID: "0694eb5e-cf26-46a4-b695-91706ed2fb13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.144843 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n" (OuterVolumeSpecName: "kube-api-access-vch5n") pod "a7523f74-b2dc-4eaa-a12e-e244510a652b" (UID: "a7523f74-b2dc-4eaa-a12e-e244510a652b"). InnerVolumeSpecName "kube-api-access-vch5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.162894 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "859fc02a-e8f2-4aec-99ce-d00f153ac581" (UID: "859fc02a-e8f2-4aec-99ce-d00f153ac581"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.167762 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp" (OuterVolumeSpecName: "kube-api-access-nksrp") pod "58e54292-9a18-4074-aa87-a9d3a8372d49" (UID: "58e54292-9a18-4074-aa87-a9d3a8372d49"). InnerVolumeSpecName "kube-api-access-nksrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.169282 4888 generic.go:334] "Generic (PLEG): container finished" podID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerID="8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34" exitCode=0 Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.169389 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerDied","Data":"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.169428 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkctm" event={"ID":"58e54292-9a18-4074-aa87-a9d3a8372d49","Type":"ContainerDied","Data":"5a48d51d1c393f1bf8ab9b257a5f8d75c324f3b2e0e3aec783864d5ba0e7de22"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.169499 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkctm" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.177178 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7" (OuterVolumeSpecName: "kube-api-access-srsj7") pod "0694eb5e-cf26-46a4-b695-91706ed2fb13" (UID: "0694eb5e-cf26-46a4-b695-91706ed2fb13"). InnerVolumeSpecName "kube-api-access-srsj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.177322 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44" (OuterVolumeSpecName: "kube-api-access-4ns44") pod "859fc02a-e8f2-4aec-99ce-d00f153ac581" (UID: "859fc02a-e8f2-4aec-99ce-d00f153ac581"). InnerVolumeSpecName "kube-api-access-4ns44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.179534 4888 generic.go:334] "Generic (PLEG): container finished" podID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerID="86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779" exitCode=0 Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.179623 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerDied","Data":"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.179690 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgpsq" event={"ID":"0694eb5e-cf26-46a4-b695-91706ed2fb13","Type":"ContainerDied","Data":"6b3b62bce8efdebb709e1cdd56f84e9b5f84fadcf29a5ce4bd7534ff7498518e"} Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.179864 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgpsq" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.215082 4888 scope.go:117] "RemoveContainer" containerID="d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.225169 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e\": container with ID starting with d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e not found: ID does not exist" containerID="d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.225226 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e"} err="failed to get container status \"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e\": rpc error: code = NotFound desc = could not find container \"d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e\": container with ID starting with d14103ce8e82cf5132627a6ac39016f187efc460562a5fbb94bad5a8e5be7c2e not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.225258 4888 scope.go:117] "RemoveContainer" containerID="faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.235352 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c084ed8-8f40-4bba-b2d8-07992d26fd35" (UID: "8c084ed8-8f40-4bba-b2d8-07992d26fd35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237309 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237335 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vch5n\" (UniqueName: \"kubernetes.io/projected/a7523f74-b2dc-4eaa-a12e-e244510a652b-kube-api-access-vch5n\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237344 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c084ed8-8f40-4bba-b2d8-07992d26fd35-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237353 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srsj7\" (UniqueName: \"kubernetes.io/projected/0694eb5e-cf26-46a4-b695-91706ed2fb13-kube-api-access-srsj7\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237365 4888 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237372 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237396 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nksrp\" (UniqueName: \"kubernetes.io/projected/58e54292-9a18-4074-aa87-a9d3a8372d49-kube-api-access-nksrp\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237406 4888 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/859fc02a-e8f2-4aec-99ce-d00f153ac581-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237426 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.237437 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ns44\" (UniqueName: \"kubernetes.io/projected/859fc02a-e8f2-4aec-99ce-d00f153ac581-kube-api-access-4ns44\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.239042 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7523f74-b2dc-4eaa-a12e-e244510a652b" (UID: "a7523f74-b2dc-4eaa-a12e-e244510a652b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.269941 4888 scope.go:117] "RemoveContainer" containerID="85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.304562 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58e54292-9a18-4074-aa87-a9d3a8372d49" (UID: "58e54292-9a18-4074-aa87-a9d3a8372d49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.322925 4888 scope.go:117] "RemoveContainer" containerID="a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.343739 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7523f74-b2dc-4eaa-a12e-e244510a652b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.343775 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e54292-9a18-4074-aa87-a9d3a8372d49-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.358074 4888 scope.go:117] "RemoveContainer" containerID="faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.362534 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nmmd"] Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.362704 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953\": container with ID starting with faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953 not found: ID does not exist" containerID="faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.363597 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953"} err="failed to get container status \"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953\": rpc error: code = NotFound desc = could not find container \"faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953\": container with ID starting with faeeceea2b001fd31a7e5e5cab9f1266a9e504aab44b61eb8c544cdf64212953 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.363633 4888 scope.go:117] "RemoveContainer" containerID="85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.365048 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920\": container with ID starting with 85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920 not found: ID does not exist" containerID="85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.365114 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920"} err="failed to get container status \"85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920\": rpc error: code = NotFound desc = could not find container \"85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920\": container with ID starting with 85ffe733b7ad72e204c9ed85cf00195c008870b42e283f52eba0d8c356344920 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: W1124 00:28:52.365092 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef22eb11_250d_44e2_a2bf_7906d7817b8b.slice/crio-81155ed69adf77f41253bac1fec646053f3a7a9cef7d45e36265cb1755e1704f WatchSource:0}: Error finding container 81155ed69adf77f41253bac1fec646053f3a7a9cef7d45e36265cb1755e1704f: Status 404 returned error can't find the container with id 81155ed69adf77f41253bac1fec646053f3a7a9cef7d45e36265cb1755e1704f Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.365155 4888 scope.go:117] "RemoveContainer" containerID="a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.365743 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685\": container with ID starting with a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685 not found: ID does not exist" containerID="a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.365787 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685"} err="failed to get container status \"a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685\": rpc error: code = NotFound desc = could not find container \"a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685\": container with ID starting with a83b5a0d5c544bb830689c521f6f9329828bec02ad561b9c0080717c83aab685 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.365825 4888 scope.go:117] "RemoveContainer" containerID="c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.389134 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0694eb5e-cf26-46a4-b695-91706ed2fb13" (UID: "0694eb5e-cf26-46a4-b695-91706ed2fb13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.394305 4888 scope.go:117] "RemoveContainer" containerID="caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.412159 4888 scope.go:117] "RemoveContainer" containerID="cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.426840 4888 scope.go:117] "RemoveContainer" containerID="c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.427217 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f\": container with ID starting with c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f not found: ID does not exist" containerID="c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.427257 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f"} err="failed to get container status \"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f\": rpc error: code = NotFound desc = could not find container \"c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f\": container with ID starting with c31bc55e08829abc13ca3c2cbfeee4cfd2735a5d0902bcbeae0f5808cba2190f not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.427285 4888 scope.go:117] "RemoveContainer" containerID="caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.427794 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4\": container with ID starting with caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4 not found: ID does not exist" containerID="caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.427975 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4"} err="failed to get container status \"caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4\": rpc error: code = NotFound desc = could not find container \"caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4\": container with ID starting with caa10e85e07c09d795ee0738b37518996eed814b7e7290ff273e962808c7aee4 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.428178 4888 scope.go:117] "RemoveContainer" containerID="cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.428973 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7\": container with ID starting with cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7 not found: ID does not exist" containerID="cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.429057 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7"} err="failed to get container status \"cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7\": rpc error: code = NotFound desc = could not find container \"cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7\": container with ID starting with cf75ee446997d01e633736f7db0e0619348353df2ad647d83b114c289e80a5e7 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.429134 4888 scope.go:117] "RemoveContainer" containerID="8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.444848 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0694eb5e-cf26-46a4-b695-91706ed2fb13-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.460357 4888 scope.go:117] "RemoveContainer" containerID="a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.464058 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.466765 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ngz8m"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.480201 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.484888 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fx6r"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.492039 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.500295 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7vpcc"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.504428 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.508923 4888 scope.go:117] "RemoveContainer" containerID="1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.513235 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mkctm"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.524299 4888 scope.go:117] "RemoveContainer" containerID="8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.524764 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34\": container with ID starting with 8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34 not found: ID does not exist" containerID="8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.524802 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34"} err="failed to get container status \"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34\": rpc error: code = NotFound desc = could not find container \"8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34\": container with ID starting with 8c30d9b3754d646558dc54ca41cb447222c411b48630f3d9d6c7dfeef9e1df34 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.524845 4888 scope.go:117] "RemoveContainer" containerID="a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.525126 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5\": container with ID starting with a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5 not found: ID does not exist" containerID="a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.525154 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5"} err="failed to get container status \"a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5\": rpc error: code = NotFound desc = could not find container \"a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5\": container with ID starting with a2dcb4d2bf77da6fd761029f098a06ccc042819f557946ed056d086c5b43ddb5 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.525196 4888 scope.go:117] "RemoveContainer" containerID="1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.525597 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4\": container with ID starting with 1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4 not found: ID does not exist" containerID="1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.525628 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4"} err="failed to get container status \"1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4\": rpc error: code = NotFound desc = could not find container \"1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4\": container with ID starting with 1edf60ce72025986cf89e126a2ca498349fcc22ebfdad5e3a56e5c98b911a6f4 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.525673 4888 scope.go:117] "RemoveContainer" containerID="86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.542739 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.547873 4888 scope.go:117] "RemoveContainer" containerID="49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.548221 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fgpsq"] Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.567587 4888 scope.go:117] "RemoveContainer" containerID="1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.581663 4888 scope.go:117] "RemoveContainer" containerID="86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.582357 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779\": container with ID starting with 86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779 not found: ID does not exist" containerID="86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.582412 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779"} err="failed to get container status \"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779\": rpc error: code = NotFound desc = could not find container \"86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779\": container with ID starting with 86298e36ce858aced1a7750b7ebe5f77fd02374db16b1928d6364e3557ebb779 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.582462 4888 scope.go:117] "RemoveContainer" containerID="49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.582948 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006\": container with ID starting with 49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006 not found: ID does not exist" containerID="49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.583008 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006"} err="failed to get container status \"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006\": rpc error: code = NotFound desc = could not find container \"49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006\": container with ID starting with 49e954ad366795c72a1ee6ad5b782e0359bbc9ef4b65848d96af0c664527d006 not found: ID does not exist" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.583040 4888 scope.go:117] "RemoveContainer" containerID="1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65" Nov 24 00:28:52 crc kubenswrapper[4888]: E1124 00:28:52.583353 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65\": container with ID starting with 1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65 not found: ID does not exist" containerID="1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65" Nov 24 00:28:52 crc kubenswrapper[4888]: I1124 00:28:52.583395 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65"} err="failed to get container status \"1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65\": rpc error: code = NotFound desc = could not find container \"1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65\": container with ID starting with 1214c9081eb5e3af9e511282dece85cec3d2a935905d494a00f5daaee3fbec65 not found: ID does not exist" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.195493 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" event={"ID":"ef22eb11-250d-44e2-a2bf-7906d7817b8b","Type":"ContainerStarted","Data":"2d98b65d71934b5e0b99b5592ff89a2ca32ff1580a3f81ce82fd3f9bdf89fe19"} Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.196086 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" event={"ID":"ef22eb11-250d-44e2-a2bf-7906d7817b8b","Type":"ContainerStarted","Data":"81155ed69adf77f41253bac1fec646053f3a7a9cef7d45e36265cb1755e1704f"} Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.196991 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.219665 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.227473 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4nmmd" podStartSLOduration=2.2274472 podStartE2EDuration="2.2274472s" podCreationTimestamp="2025-11-24 00:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:28:53.225695031 +0000 UTC m=+235.808379085" watchObservedRunningTime="2025-11-24 00:28:53.2274472 +0000 UTC m=+235.810131244" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512051 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4zc5l"] Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512357 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512376 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512389 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512399 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512412 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512423 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512433 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512441 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512452 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512460 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512475 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512484 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512494 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512503 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512515 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512523 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="extract-content" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512531 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512538 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512547 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512555 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512565 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512572 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512583 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512591 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: E1124 00:28:53.512602 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512610 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="extract-utilities" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512725 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512745 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" containerName="marketplace-operator" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512753 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512764 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.512773 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" containerName="registry-server" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.513724 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.515826 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.542716 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zc5l"] Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.560078 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-catalog-content\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.560200 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-utilities\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.560300 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p27j\" (UniqueName: \"kubernetes.io/projected/042dca76-9304-4100-87ff-a67fa19b4674-kube-api-access-6p27j\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.661671 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p27j\" (UniqueName: \"kubernetes.io/projected/042dca76-9304-4100-87ff-a67fa19b4674-kube-api-access-6p27j\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.661758 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-catalog-content\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.661840 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-utilities\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.662363 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-catalog-content\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.662503 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042dca76-9304-4100-87ff-a67fa19b4674-utilities\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.683155 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p27j\" (UniqueName: \"kubernetes.io/projected/042dca76-9304-4100-87ff-a67fa19b4674-kube-api-access-6p27j\") pod \"redhat-marketplace-4zc5l\" (UID: \"042dca76-9304-4100-87ff-a67fa19b4674\") " pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:53 crc kubenswrapper[4888]: I1124 00:28:53.843956 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.058691 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zc5l"] Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.117617 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qlhkx"] Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.119057 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.119966 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qlhkx"] Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.122014 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.167885 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-utilities\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.167985 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-catalog-content\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.168017 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxzld\" (UniqueName: \"kubernetes.io/projected/cab99235-2ee5-487d-be36-3d78fff69946-kube-api-access-rxzld\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.211217 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zc5l" event={"ID":"042dca76-9304-4100-87ff-a67fa19b4674","Type":"ContainerStarted","Data":"08da8d65d12d306d8277ecab68a057d9929ac0fa5a7a937302b0f23d8ee2ed79"} Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.252141 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0694eb5e-cf26-46a4-b695-91706ed2fb13" path="/var/lib/kubelet/pods/0694eb5e-cf26-46a4-b695-91706ed2fb13/volumes" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.253181 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58e54292-9a18-4074-aa87-a9d3a8372d49" path="/var/lib/kubelet/pods/58e54292-9a18-4074-aa87-a9d3a8372d49/volumes" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.254008 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859fc02a-e8f2-4aec-99ce-d00f153ac581" path="/var/lib/kubelet/pods/859fc02a-e8f2-4aec-99ce-d00f153ac581/volumes" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.255639 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c084ed8-8f40-4bba-b2d8-07992d26fd35" path="/var/lib/kubelet/pods/8c084ed8-8f40-4bba-b2d8-07992d26fd35/volumes" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.256305 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7523f74-b2dc-4eaa-a12e-e244510a652b" path="/var/lib/kubelet/pods/a7523f74-b2dc-4eaa-a12e-e244510a652b/volumes" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.269798 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-utilities\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.269899 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-catalog-content\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.269929 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxzld\" (UniqueName: \"kubernetes.io/projected/cab99235-2ee5-487d-be36-3d78fff69946-kube-api-access-rxzld\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.270289 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-utilities\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.270713 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab99235-2ee5-487d-be36-3d78fff69946-catalog-content\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.293803 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxzld\" (UniqueName: \"kubernetes.io/projected/cab99235-2ee5-487d-be36-3d78fff69946-kube-api-access-rxzld\") pod \"certified-operators-qlhkx\" (UID: \"cab99235-2ee5-487d-be36-3d78fff69946\") " pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.463676 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:28:54 crc kubenswrapper[4888]: I1124 00:28:54.706502 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qlhkx"] Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.220082 4888 generic.go:334] "Generic (PLEG): container finished" podID="042dca76-9304-4100-87ff-a67fa19b4674" containerID="be12df03a8d711d1dcd3f94216681357c800ccfef83f6fc735e42d00272f0493" exitCode=0 Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.220192 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zc5l" event={"ID":"042dca76-9304-4100-87ff-a67fa19b4674","Type":"ContainerDied","Data":"be12df03a8d711d1dcd3f94216681357c800ccfef83f6fc735e42d00272f0493"} Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.222109 4888 generic.go:334] "Generic (PLEG): container finished" podID="cab99235-2ee5-487d-be36-3d78fff69946" containerID="603d3ca836b028fb2c0cc4797f06ef3a372dbc8daf40fab6a2471106325acea3" exitCode=0 Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.222763 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhkx" event={"ID":"cab99235-2ee5-487d-be36-3d78fff69946","Type":"ContainerDied","Data":"603d3ca836b028fb2c0cc4797f06ef3a372dbc8daf40fab6a2471106325acea3"} Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.222784 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhkx" event={"ID":"cab99235-2ee5-487d-be36-3d78fff69946","Type":"ContainerStarted","Data":"bfe672f7518ce5c363ca79d5e1959cc6054899b6b1363bdaa768dd58d2209331"} Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.907776 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v97rx"] Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.910602 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.915418 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 00:28:55 crc kubenswrapper[4888]: I1124 00:28:55.926568 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v97rx"] Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.009728 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-catalog-content\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.009825 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-utilities\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.009879 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbff5\" (UniqueName: \"kubernetes.io/projected/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-kube-api-access-dbff5\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.111133 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbff5\" (UniqueName: \"kubernetes.io/projected/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-kube-api-access-dbff5\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.111685 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-catalog-content\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.111746 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-utilities\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.112518 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-catalog-content\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.112542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-utilities\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.155975 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbff5\" (UniqueName: \"kubernetes.io/projected/d48aee78-3eb2-439e-9ae4-6d56a796ad1c-kube-api-access-dbff5\") pod \"redhat-operators-v97rx\" (UID: \"d48aee78-3eb2-439e-9ae4-6d56a796ad1c\") " pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.230118 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhkx" event={"ID":"cab99235-2ee5-487d-be36-3d78fff69946","Type":"ContainerStarted","Data":"e6b871cec2307b0f1076959bd6bf7928f0ec0c95edab4d02fe09078157a58ddc"} Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.249130 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.462222 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v97rx"] Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.521896 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p29pb"] Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.524318 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.527522 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.531202 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p29pb"] Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.620917 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-utilities\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.621017 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh5fq\" (UniqueName: \"kubernetes.io/projected/6edb2361-460b-4aa7-8d4c-0338d630b1ca-kube-api-access-jh5fq\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.621179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-catalog-content\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.722882 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh5fq\" (UniqueName: \"kubernetes.io/projected/6edb2361-460b-4aa7-8d4c-0338d630b1ca-kube-api-access-jh5fq\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.722936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-catalog-content\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.722990 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-utilities\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.723842 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-utilities\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.725654 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6edb2361-460b-4aa7-8d4c-0338d630b1ca-catalog-content\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.747533 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh5fq\" (UniqueName: \"kubernetes.io/projected/6edb2361-460b-4aa7-8d4c-0338d630b1ca-kube-api-access-jh5fq\") pod \"community-operators-p29pb\" (UID: \"6edb2361-460b-4aa7-8d4c-0338d630b1ca\") " pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:56 crc kubenswrapper[4888]: I1124 00:28:56.866702 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.120842 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p29pb"] Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.240287 4888 generic.go:334] "Generic (PLEG): container finished" podID="d48aee78-3eb2-439e-9ae4-6d56a796ad1c" containerID="6b2a1a01c6959a4c3255a76cacf1d8956bea940a3cc282ebeaea8582b2c4dfa2" exitCode=0 Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.240394 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v97rx" event={"ID":"d48aee78-3eb2-439e-9ae4-6d56a796ad1c","Type":"ContainerDied","Data":"6b2a1a01c6959a4c3255a76cacf1d8956bea940a3cc282ebeaea8582b2c4dfa2"} Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.240458 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v97rx" event={"ID":"d48aee78-3eb2-439e-9ae4-6d56a796ad1c","Type":"ContainerStarted","Data":"bb9c1f1c9c5f18a6a631e35ecadf419d15792b65cf192a3344d7d7db75c5c521"} Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.243636 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p29pb" event={"ID":"6edb2361-460b-4aa7-8d4c-0338d630b1ca","Type":"ContainerStarted","Data":"b652eda4711ebaef8c7fdbf4179744f8763a7866ce155230ea933f39aecd1fa7"} Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.246612 4888 generic.go:334] "Generic (PLEG): container finished" podID="cab99235-2ee5-487d-be36-3d78fff69946" containerID="e6b871cec2307b0f1076959bd6bf7928f0ec0c95edab4d02fe09078157a58ddc" exitCode=0 Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.246712 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhkx" event={"ID":"cab99235-2ee5-487d-be36-3d78fff69946","Type":"ContainerDied","Data":"e6b871cec2307b0f1076959bd6bf7928f0ec0c95edab4d02fe09078157a58ddc"} Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.255691 4888 generic.go:334] "Generic (PLEG): container finished" podID="042dca76-9304-4100-87ff-a67fa19b4674" containerID="6ac2bed2ab2770b0be570f061d998787e0494ba455f789a86cda6864967cff7c" exitCode=0 Nov 24 00:28:57 crc kubenswrapper[4888]: I1124 00:28:57.255742 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zc5l" event={"ID":"042dca76-9304-4100-87ff-a67fa19b4674","Type":"ContainerDied","Data":"6ac2bed2ab2770b0be570f061d998787e0494ba455f789a86cda6864967cff7c"} Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.266160 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhkx" event={"ID":"cab99235-2ee5-487d-be36-3d78fff69946","Type":"ContainerStarted","Data":"9725948ad06ff0cbf5de7e36d5a0385417a80fc4b8f68f88e701d7ceb39ffbf7"} Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.281476 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zc5l" event={"ID":"042dca76-9304-4100-87ff-a67fa19b4674","Type":"ContainerStarted","Data":"3112db3215a414311371c3abb21f4f5997ae37d026a13492a62447593665ba09"} Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.284491 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v97rx" event={"ID":"d48aee78-3eb2-439e-9ae4-6d56a796ad1c","Type":"ContainerStarted","Data":"3ba4e66d4094f6be0c8a50c2de8bbc077a9116241bd6fcf19732ea6672c0f310"} Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.297454 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qlhkx" podStartSLOduration=1.839598377 podStartE2EDuration="4.297425425s" podCreationTimestamp="2025-11-24 00:28:54 +0000 UTC" firstStartedPulling="2025-11-24 00:28:55.223969171 +0000 UTC m=+237.806653295" lastFinishedPulling="2025-11-24 00:28:57.681796299 +0000 UTC m=+240.264480343" observedRunningTime="2025-11-24 00:28:58.28877137 +0000 UTC m=+240.871455414" watchObservedRunningTime="2025-11-24 00:28:58.297425425 +0000 UTC m=+240.880109469" Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.298503 4888 generic.go:334] "Generic (PLEG): container finished" podID="6edb2361-460b-4aa7-8d4c-0338d630b1ca" containerID="0f10219ac5b8ff2111558d4a4e4ae5d898eb15f57e12feb499a0b20a39118340" exitCode=0 Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.298540 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p29pb" event={"ID":"6edb2361-460b-4aa7-8d4c-0338d630b1ca","Type":"ContainerDied","Data":"0f10219ac5b8ff2111558d4a4e4ae5d898eb15f57e12feb499a0b20a39118340"} Nov 24 00:28:58 crc kubenswrapper[4888]: I1124 00:28:58.313755 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4zc5l" podStartSLOduration=2.845002971 podStartE2EDuration="5.31373457s" podCreationTimestamp="2025-11-24 00:28:53 +0000 UTC" firstStartedPulling="2025-11-24 00:28:55.222881644 +0000 UTC m=+237.805565688" lastFinishedPulling="2025-11-24 00:28:57.691613203 +0000 UTC m=+240.274297287" observedRunningTime="2025-11-24 00:28:58.311605777 +0000 UTC m=+240.894289831" watchObservedRunningTime="2025-11-24 00:28:58.31373457 +0000 UTC m=+240.896418614" Nov 24 00:28:59 crc kubenswrapper[4888]: I1124 00:28:59.307214 4888 generic.go:334] "Generic (PLEG): container finished" podID="6edb2361-460b-4aa7-8d4c-0338d630b1ca" containerID="4a38f605d9a9f2414f5876ef04b858dfb526fae098138e3343547b580c3e3b73" exitCode=0 Nov 24 00:28:59 crc kubenswrapper[4888]: I1124 00:28:59.307299 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p29pb" event={"ID":"6edb2361-460b-4aa7-8d4c-0338d630b1ca","Type":"ContainerDied","Data":"4a38f605d9a9f2414f5876ef04b858dfb526fae098138e3343547b580c3e3b73"} Nov 24 00:28:59 crc kubenswrapper[4888]: I1124 00:28:59.312950 4888 generic.go:334] "Generic (PLEG): container finished" podID="d48aee78-3eb2-439e-9ae4-6d56a796ad1c" containerID="3ba4e66d4094f6be0c8a50c2de8bbc077a9116241bd6fcf19732ea6672c0f310" exitCode=0 Nov 24 00:28:59 crc kubenswrapper[4888]: I1124 00:28:59.313048 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v97rx" event={"ID":"d48aee78-3eb2-439e-9ae4-6d56a796ad1c","Type":"ContainerDied","Data":"3ba4e66d4094f6be0c8a50c2de8bbc077a9116241bd6fcf19732ea6672c0f310"} Nov 24 00:29:01 crc kubenswrapper[4888]: I1124 00:29:01.327120 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v97rx" event={"ID":"d48aee78-3eb2-439e-9ae4-6d56a796ad1c","Type":"ContainerStarted","Data":"cee544599756d049b06f27da0ba7f9d7d33b018106ba3efa8feb0f99e3795ea1"} Nov 24 00:29:01 crc kubenswrapper[4888]: I1124 00:29:01.337064 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p29pb" event={"ID":"6edb2361-460b-4aa7-8d4c-0338d630b1ca","Type":"ContainerStarted","Data":"a0ca3343c7089d1b6bf26230d9a80beeb5444d4ebba1c0730a70c8a531adb7a5"} Nov 24 00:29:01 crc kubenswrapper[4888]: I1124 00:29:01.355424 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v97rx" podStartSLOduration=3.8679014560000002 podStartE2EDuration="6.355396123s" podCreationTimestamp="2025-11-24 00:28:55 +0000 UTC" firstStartedPulling="2025-11-24 00:28:57.246667383 +0000 UTC m=+239.829351427" lastFinishedPulling="2025-11-24 00:28:59.73416204 +0000 UTC m=+242.316846094" observedRunningTime="2025-11-24 00:29:01.349517603 +0000 UTC m=+243.932201657" watchObservedRunningTime="2025-11-24 00:29:01.355396123 +0000 UTC m=+243.938080197" Nov 24 00:29:01 crc kubenswrapper[4888]: I1124 00:29:01.376016 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p29pb" podStartSLOduration=3.815043762 podStartE2EDuration="5.375991964s" podCreationTimestamp="2025-11-24 00:28:56 +0000 UTC" firstStartedPulling="2025-11-24 00:28:58.299974221 +0000 UTC m=+240.882658265" lastFinishedPulling="2025-11-24 00:28:59.860922423 +0000 UTC m=+242.443606467" observedRunningTime="2025-11-24 00:29:01.373170228 +0000 UTC m=+243.955854272" watchObservedRunningTime="2025-11-24 00:29:01.375991964 +0000 UTC m=+243.958676018" Nov 24 00:29:03 crc kubenswrapper[4888]: I1124 00:29:03.844443 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:29:03 crc kubenswrapper[4888]: I1124 00:29:03.845804 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:29:03 crc kubenswrapper[4888]: I1124 00:29:03.904069 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:29:04 crc kubenswrapper[4888]: I1124 00:29:04.397422 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4zc5l" Nov 24 00:29:04 crc kubenswrapper[4888]: I1124 00:29:04.464974 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:29:04 crc kubenswrapper[4888]: I1124 00:29:04.465037 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:29:04 crc kubenswrapper[4888]: I1124 00:29:04.510430 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:29:05 crc kubenswrapper[4888]: I1124 00:29:05.417841 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qlhkx" Nov 24 00:29:06 crc kubenswrapper[4888]: I1124 00:29:06.251761 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:29:06 crc kubenswrapper[4888]: I1124 00:29:06.251820 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:29:06 crc kubenswrapper[4888]: I1124 00:29:06.867241 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:29:06 crc kubenswrapper[4888]: I1124 00:29:06.867318 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:29:06 crc kubenswrapper[4888]: I1124 00:29:06.923092 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:29:07 crc kubenswrapper[4888]: I1124 00:29:07.310028 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v97rx" podUID="d48aee78-3eb2-439e-9ae4-6d56a796ad1c" containerName="registry-server" probeResult="failure" output=< Nov 24 00:29:07 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:29:07 crc kubenswrapper[4888]: > Nov 24 00:29:07 crc kubenswrapper[4888]: I1124 00:29:07.421419 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p29pb" Nov 24 00:29:16 crc kubenswrapper[4888]: I1124 00:29:16.300286 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:29:16 crc kubenswrapper[4888]: I1124 00:29:16.355226 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v97rx" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.167712 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q"] Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.175392 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.179494 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q"] Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.180268 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.180713 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.275649 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnf8\" (UniqueName: \"kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.275831 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.275870 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.378106 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnf8\" (UniqueName: \"kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.378709 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.378733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.379778 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.388088 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.400859 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnf8\" (UniqueName: \"kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8\") pod \"collect-profiles-29399070-pvx4q\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.498509 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.716772 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q"] Nov 24 00:30:00 crc kubenswrapper[4888]: I1124 00:30:00.748593 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" event={"ID":"0282dd1d-4a52-4986-a79c-01d365dab893","Type":"ContainerStarted","Data":"99ddc8aec57911123554145c900f70e2efdb1ec545702bf7cdb840e23e37b0f5"} Nov 24 00:30:01 crc kubenswrapper[4888]: I1124 00:30:01.756394 4888 generic.go:334] "Generic (PLEG): container finished" podID="0282dd1d-4a52-4986-a79c-01d365dab893" containerID="7a4a5bec4038ef1dfcec7a467b296167de8a934cb72f2cc9b9f80c6829a521f6" exitCode=0 Nov 24 00:30:01 crc kubenswrapper[4888]: I1124 00:30:01.756453 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" event={"ID":"0282dd1d-4a52-4986-a79c-01d365dab893","Type":"ContainerDied","Data":"7a4a5bec4038ef1dfcec7a467b296167de8a934cb72f2cc9b9f80c6829a521f6"} Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.053209 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.119241 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume\") pod \"0282dd1d-4a52-4986-a79c-01d365dab893\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.119718 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume\") pod \"0282dd1d-4a52-4986-a79c-01d365dab893\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.120153 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfnf8\" (UniqueName: \"kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8\") pod \"0282dd1d-4a52-4986-a79c-01d365dab893\" (UID: \"0282dd1d-4a52-4986-a79c-01d365dab893\") " Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.120543 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume" (OuterVolumeSpecName: "config-volume") pod "0282dd1d-4a52-4986-a79c-01d365dab893" (UID: "0282dd1d-4a52-4986-a79c-01d365dab893"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.121268 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0282dd1d-4a52-4986-a79c-01d365dab893-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.126702 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8" (OuterVolumeSpecName: "kube-api-access-gfnf8") pod "0282dd1d-4a52-4986-a79c-01d365dab893" (UID: "0282dd1d-4a52-4986-a79c-01d365dab893"). InnerVolumeSpecName "kube-api-access-gfnf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.130086 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0282dd1d-4a52-4986-a79c-01d365dab893" (UID: "0282dd1d-4a52-4986-a79c-01d365dab893"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.222340 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0282dd1d-4a52-4986-a79c-01d365dab893-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.222381 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfnf8\" (UniqueName: \"kubernetes.io/projected/0282dd1d-4a52-4986-a79c-01d365dab893-kube-api-access-gfnf8\") on node \"crc\" DevicePath \"\"" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.772182 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" event={"ID":"0282dd1d-4a52-4986-a79c-01d365dab893","Type":"ContainerDied","Data":"99ddc8aec57911123554145c900f70e2efdb1ec545702bf7cdb840e23e37b0f5"} Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.772644 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99ddc8aec57911123554145c900f70e2efdb1ec545702bf7cdb840e23e37b0f5" Nov 24 00:30:03 crc kubenswrapper[4888]: I1124 00:30:03.772279 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q" Nov 24 00:30:23 crc kubenswrapper[4888]: I1124 00:30:23.591948 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:30:23 crc kubenswrapper[4888]: I1124 00:30:23.592970 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:30:53 crc kubenswrapper[4888]: I1124 00:30:53.591879 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:30:53 crc kubenswrapper[4888]: I1124 00:30:53.592960 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:31:23 crc kubenswrapper[4888]: I1124 00:31:23.592247 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:31:23 crc kubenswrapper[4888]: I1124 00:31:23.593170 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:31:23 crc kubenswrapper[4888]: I1124 00:31:23.593253 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:31:23 crc kubenswrapper[4888]: I1124 00:31:23.594273 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:31:23 crc kubenswrapper[4888]: I1124 00:31:23.594386 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4" gracePeriod=600 Nov 24 00:31:24 crc kubenswrapper[4888]: I1124 00:31:24.301271 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4" exitCode=0 Nov 24 00:31:24 crc kubenswrapper[4888]: I1124 00:31:24.301373 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4"} Nov 24 00:31:24 crc kubenswrapper[4888]: I1124 00:31:24.302317 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde"} Nov 24 00:31:24 crc kubenswrapper[4888]: I1124 00:31:24.302359 4888 scope.go:117] "RemoveContainer" containerID="451eb1aa31142da8c0cfa5488a310cff3595ae4b0916bd9119899a19528701e7" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.760789 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9dnz"] Nov 24 00:32:09 crc kubenswrapper[4888]: E1124 00:32:09.765962 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0282dd1d-4a52-4986-a79c-01d365dab893" containerName="collect-profiles" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.765983 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0282dd1d-4a52-4986-a79c-01d365dab893" containerName="collect-profiles" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.766102 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0282dd1d-4a52-4986-a79c-01d365dab893" containerName="collect-profiles" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.766558 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.773104 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9dnz"] Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897603 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897680 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-bound-sa-token\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897804 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897900 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-tls\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.897936 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w78jd\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-kube-api-access-w78jd\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.898011 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-trusted-ca\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.898056 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-certificates\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.925743 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:09 crc kubenswrapper[4888]: I1124 00:32:09.999749 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000180 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-bound-sa-token\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000324 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000455 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-tls\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w78jd\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-kube-api-access-w78jd\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000654 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-trusted-ca\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.000752 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-certificates\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.001654 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.002985 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-certificates\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.003089 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-trusted-ca\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.008677 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-registry-tls\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.008748 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.020374 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-bound-sa-token\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.027422 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w78jd\" (UniqueName: \"kubernetes.io/projected/76b5a2a8-b8fc-4cd0-9790-1bf703966f83-kube-api-access-w78jd\") pod \"image-registry-66df7c8f76-d9dnz\" (UID: \"76b5a2a8-b8fc-4cd0-9790-1bf703966f83\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.089428 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.319946 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9dnz"] Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.638985 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" event={"ID":"76b5a2a8-b8fc-4cd0-9790-1bf703966f83","Type":"ContainerStarted","Data":"40ad086808c953e3ab5d73bba374926f4eaea37b44c21097a32eb260a24ed2dd"} Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.639575 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.639586 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" event={"ID":"76b5a2a8-b8fc-4cd0-9790-1bf703966f83","Type":"ContainerStarted","Data":"48d82ece11b53cb544b064f89b4bc7f9220254f2ff5e6217e1f03e9e68bfe90d"} Nov 24 00:32:10 crc kubenswrapper[4888]: I1124 00:32:10.661918 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" podStartSLOduration=1.661896976 podStartE2EDuration="1.661896976s" podCreationTimestamp="2025-11-24 00:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:32:10.659591068 +0000 UTC m=+433.242275112" watchObservedRunningTime="2025-11-24 00:32:10.661896976 +0000 UTC m=+433.244581020" Nov 24 00:32:30 crc kubenswrapper[4888]: I1124 00:32:30.094768 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-d9dnz" Nov 24 00:32:30 crc kubenswrapper[4888]: I1124 00:32:30.151699 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.198562 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" podUID="1a2cedaf-0125-42ee-969b-59699128ebcc" containerName="registry" containerID="cri-o://1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c" gracePeriod=30 Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.598162 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709062 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709180 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709209 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-png9r\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709243 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709276 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709307 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709361 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.709383 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates\") pod \"1a2cedaf-0125-42ee-969b-59699128ebcc\" (UID: \"1a2cedaf-0125-42ee-969b-59699128ebcc\") " Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.710913 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.710935 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.718110 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.718262 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.719038 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.723654 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r" (OuterVolumeSpecName: "kube-api-access-png9r") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "kube-api-access-png9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.725227 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.726676 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1a2cedaf-0125-42ee-969b-59699128ebcc" (UID: "1a2cedaf-0125-42ee-969b-59699128ebcc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.790711 4888 generic.go:334] "Generic (PLEG): container finished" podID="1a2cedaf-0125-42ee-969b-59699128ebcc" containerID="1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c" exitCode=0 Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.790769 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" event={"ID":"1a2cedaf-0125-42ee-969b-59699128ebcc","Type":"ContainerDied","Data":"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c"} Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.790861 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" event={"ID":"1a2cedaf-0125-42ee-969b-59699128ebcc","Type":"ContainerDied","Data":"364d0be16ea1cd06e68ad8e1db0646ca7cf6719bf644023158c5a9bb620c0cda"} Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.790878 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d8t6" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.790898 4888 scope.go:117] "RemoveContainer" containerID="1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810856 4888 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810904 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-png9r\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-kube-api-access-png9r\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810921 4888 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a2cedaf-0125-42ee-969b-59699128ebcc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810936 4888 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810949 4888 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a2cedaf-0125-42ee-969b-59699128ebcc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810961 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.810973 4888 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a2cedaf-0125-42ee-969b-59699128ebcc-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.812340 4888 scope.go:117] "RemoveContainer" containerID="1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c" Nov 24 00:32:55 crc kubenswrapper[4888]: E1124 00:32:55.812945 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c\": container with ID starting with 1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c not found: ID does not exist" containerID="1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.812984 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c"} err="failed to get container status \"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c\": rpc error: code = NotFound desc = could not find container \"1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c\": container with ID starting with 1bed9044ecf4028cef6dccebe38a6f55d4fbb6d5e56ae09ef07b091a0b39a09c not found: ID does not exist" Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.835735 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:32:55 crc kubenswrapper[4888]: I1124 00:32:55.841335 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d8t6"] Nov 24 00:32:56 crc kubenswrapper[4888]: I1124 00:32:56.257315 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a2cedaf-0125-42ee-969b-59699128ebcc" path="/var/lib/kubelet/pods/1a2cedaf-0125-42ee-969b-59699128ebcc/volumes" Nov 24 00:33:23 crc kubenswrapper[4888]: I1124 00:33:23.591466 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:33:23 crc kubenswrapper[4888]: I1124 00:33:23.592391 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:33:53 crc kubenswrapper[4888]: I1124 00:33:53.591515 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:33:53 crc kubenswrapper[4888]: I1124 00:33:53.592609 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.586878 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm"] Nov 24 00:34:10 crc kubenswrapper[4888]: E1124 00:34:10.587831 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2cedaf-0125-42ee-969b-59699128ebcc" containerName="registry" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.587850 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2cedaf-0125-42ee-969b-59699128ebcc" containerName="registry" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.587986 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2cedaf-0125-42ee-969b-59699128ebcc" containerName="registry" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.588987 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.592015 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.597535 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm"] Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.691891 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.692007 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncqwj\" (UniqueName: \"kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.692060 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.793341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncqwj\" (UniqueName: \"kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.793442 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.793487 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.794077 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.794090 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.811883 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncqwj\" (UniqueName: \"kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:10 crc kubenswrapper[4888]: I1124 00:34:10.946765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:11 crc kubenswrapper[4888]: I1124 00:34:11.165972 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm"] Nov 24 00:34:11 crc kubenswrapper[4888]: I1124 00:34:11.318285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" event={"ID":"04e91586-5c1b-490b-ac59-7f0f162623cd","Type":"ContainerStarted","Data":"d74b3123567927fdedbd69861696cee21760705b3ff0ff26553a24af2db328b6"} Nov 24 00:34:12 crc kubenswrapper[4888]: I1124 00:34:12.325785 4888 generic.go:334] "Generic (PLEG): container finished" podID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerID="84f1481ee8a8ca069fda032fc213665db2b98b94fd339c1d0f9c2cf48405f248" exitCode=0 Nov 24 00:34:12 crc kubenswrapper[4888]: I1124 00:34:12.325898 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" event={"ID":"04e91586-5c1b-490b-ac59-7f0f162623cd","Type":"ContainerDied","Data":"84f1481ee8a8ca069fda032fc213665db2b98b94fd339c1d0f9c2cf48405f248"} Nov 24 00:34:12 crc kubenswrapper[4888]: I1124 00:34:12.329152 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:34:14 crc kubenswrapper[4888]: I1124 00:34:14.347734 4888 generic.go:334] "Generic (PLEG): container finished" podID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerID="228945c5978572a62121b4b816ac1dd76dd17387bd975ec25f6dcc4bbb770e0d" exitCode=0 Nov 24 00:34:14 crc kubenswrapper[4888]: I1124 00:34:14.347868 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" event={"ID":"04e91586-5c1b-490b-ac59-7f0f162623cd","Type":"ContainerDied","Data":"228945c5978572a62121b4b816ac1dd76dd17387bd975ec25f6dcc4bbb770e0d"} Nov 24 00:34:15 crc kubenswrapper[4888]: I1124 00:34:15.360493 4888 generic.go:334] "Generic (PLEG): container finished" podID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerID="3a19b99a91830739592f49ef9fe264f367f1750744f098485d6959f92893d36c" exitCode=0 Nov 24 00:34:15 crc kubenswrapper[4888]: I1124 00:34:15.360702 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" event={"ID":"04e91586-5c1b-490b-ac59-7f0f162623cd","Type":"ContainerDied","Data":"3a19b99a91830739592f49ef9fe264f367f1750744f098485d6959f92893d36c"} Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.619088 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.787125 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util\") pod \"04e91586-5c1b-490b-ac59-7f0f162623cd\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.787269 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle\") pod \"04e91586-5c1b-490b-ac59-7f0f162623cd\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.787363 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncqwj\" (UniqueName: \"kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj\") pod \"04e91586-5c1b-490b-ac59-7f0f162623cd\" (UID: \"04e91586-5c1b-490b-ac59-7f0f162623cd\") " Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.790541 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle" (OuterVolumeSpecName: "bundle") pod "04e91586-5c1b-490b-ac59-7f0f162623cd" (UID: "04e91586-5c1b-490b-ac59-7f0f162623cd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.794561 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj" (OuterVolumeSpecName: "kube-api-access-ncqwj") pod "04e91586-5c1b-490b-ac59-7f0f162623cd" (UID: "04e91586-5c1b-490b-ac59-7f0f162623cd"). InnerVolumeSpecName "kube-api-access-ncqwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.815100 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util" (OuterVolumeSpecName: "util") pod "04e91586-5c1b-490b-ac59-7f0f162623cd" (UID: "04e91586-5c1b-490b-ac59-7f0f162623cd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.891208 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.891257 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04e91586-5c1b-490b-ac59-7f0f162623cd-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:16 crc kubenswrapper[4888]: I1124 00:34:16.891274 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncqwj\" (UniqueName: \"kubernetes.io/projected/04e91586-5c1b-490b-ac59-7f0f162623cd-kube-api-access-ncqwj\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:17 crc kubenswrapper[4888]: I1124 00:34:17.380419 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" event={"ID":"04e91586-5c1b-490b-ac59-7f0f162623cd","Type":"ContainerDied","Data":"d74b3123567927fdedbd69861696cee21760705b3ff0ff26553a24af2db328b6"} Nov 24 00:34:17 crc kubenswrapper[4888]: I1124 00:34:17.380507 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d74b3123567927fdedbd69861696cee21760705b3ff0ff26553a24af2db328b6" Nov 24 00:34:17 crc kubenswrapper[4888]: I1124 00:34:17.380644 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm" Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.666614 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6j5lq"] Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668059 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-controller" containerID="cri-o://e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668700 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="sbdb" containerID="cri-o://bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668773 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="nbdb" containerID="cri-o://20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668864 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="northd" containerID="cri-o://5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668927 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.668991 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-node" containerID="cri-o://eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.669056 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-acl-logging" containerID="cri-o://9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" gracePeriod=30 Nov 24 00:34:21 crc kubenswrapper[4888]: I1124 00:34:21.713794 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" containerID="cri-o://f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" gracePeriod=30 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.044121 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/3.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.046904 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovn-acl-logging/0.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.047494 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovn-controller/0.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.048015 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.117917 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6pmlv"] Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118210 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118232 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118244 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="util" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118251 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="util" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118261 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-node" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118269 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-node" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118282 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118289 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118298 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kubecfg-setup" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118305 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kubecfg-setup" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118319 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118326 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118337 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118345 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118359 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118370 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118380 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-acl-logging" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118388 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-acl-logging" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118399 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="sbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118405 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="sbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118415 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="northd" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118423 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="northd" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118435 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="nbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118443 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="nbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118543 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="extract" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118553 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="extract" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118564 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="pull" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118571 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="pull" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118697 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e91586-5c1b-490b-ac59-7f0f162623cd" containerName="extract" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118712 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="nbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118727 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="northd" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118735 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118745 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118754 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-node" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118762 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-acl-logging" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118769 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118778 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118787 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovn-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118794 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="sbdb" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118802 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118828 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.118937 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.118947 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.119188 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.119197 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerName="ovnkube-controller" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.121025 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170039 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170043 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170148 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170178 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170209 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz76r\" (UniqueName: \"kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170230 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170252 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170303 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170303 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170377 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170784 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170850 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170868 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170880 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170890 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170912 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170925 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170990 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170933 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170950 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.170956 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171066 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171119 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket" (OuterVolumeSpecName: "log-socket") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171220 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171225 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171450 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171093 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171566 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171589 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.171774 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log" (OuterVolumeSpecName: "node-log") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173021 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173122 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173252 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173360 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173453 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash" (OuterVolumeSpecName: "host-slash") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173559 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173558 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes\") pod \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\" (UID: \"2d662e59-55e9-45fa-870d-ad8c51a7a0ef\") " Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173673 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-log-socket\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173704 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173724 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-script-lib\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173790 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-slash\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173845 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-bin\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173878 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2pmm\" (UniqueName: \"kubernetes.io/projected/0f1d570f-0b3b-4553-a9df-a223f9c4c109-kube-api-access-j2pmm\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173912 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-netd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173955 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovn-node-metrics-cert\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.173994 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-var-lib-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174033 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-kubelet\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174054 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174086 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-etc-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174105 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-systemd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174159 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-netns\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174185 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-ovn\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174203 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-config\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174230 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-node-log\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174249 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-systemd-units\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174272 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-env-overrides\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174290 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174347 4888 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174360 4888 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174371 4888 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174382 4888 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174394 4888 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174403 4888 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174412 4888 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174422 4888 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174431 4888 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174439 4888 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174451 4888 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174461 4888 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174469 4888 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174478 4888 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174486 4888 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174495 4888 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.174504 4888 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.178158 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.178729 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r" (OuterVolumeSpecName: "kube-api-access-mz76r") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "kube-api-access-mz76r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.185669 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2d662e59-55e9-45fa-870d-ad8c51a7a0ef" (UID: "2d662e59-55e9-45fa-870d-ad8c51a7a0ef"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275682 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-var-lib-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275773 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-kubelet\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275898 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-var-lib-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275957 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275924 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-kubelet\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.275960 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276059 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-etc-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276081 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-systemd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276135 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-netns\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276139 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-etc-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276182 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-ovn\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276205 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-config\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276225 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-node-log\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-systemd-units\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276249 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-systemd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276280 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-env-overrides\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276305 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276330 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-node-log\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276344 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-log-socket\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276361 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276378 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-ovn\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276384 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-script-lib\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276444 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-slash\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276480 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-bin\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276535 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2pmm\" (UniqueName: \"kubernetes.io/projected/0f1d570f-0b3b-4553-a9df-a223f9c4c109-kube-api-access-j2pmm\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276576 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-netd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276635 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovn-node-metrics-cert\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276697 4888 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276720 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276740 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz76r\" (UniqueName: \"kubernetes.io/projected/2d662e59-55e9-45fa-870d-ad8c51a7a0ef-kube-api-access-mz76r\") on node \"crc\" DevicePath \"\"" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276865 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-run-openvswitch\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276225 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-run-netns\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.276919 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-systemd-units\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277138 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-script-lib\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277280 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-netd\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277433 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-env-overrides\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277500 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-cni-bin\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277540 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-log-socket\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277566 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.277443 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f1d570f-0b3b-4553-a9df-a223f9c4c109-host-slash\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.278014 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovnkube-config\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.282360 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f1d570f-0b3b-4553-a9df-a223f9c4c109-ovn-node-metrics-cert\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.303702 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2pmm\" (UniqueName: \"kubernetes.io/projected/0f1d570f-0b3b-4553-a9df-a223f9c4c109-kube-api-access-j2pmm\") pod \"ovnkube-node-6pmlv\" (UID: \"0f1d570f-0b3b-4553-a9df-a223f9c4c109\") " pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.416520 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/2.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.416997 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/1.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.417048 4888 generic.go:334] "Generic (PLEG): container finished" podID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" containerID="2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799" exitCode=2 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.417115 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerDied","Data":"2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.417159 4888 scope.go:117] "RemoveContainer" containerID="16104d5ac150679bfdb0384078e5cf367f9d6d8406b466db0ad8b9aa0aa6adc7" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.418004 4888 scope.go:117] "RemoveContainer" containerID="2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.418285 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-zbfxr_openshift-multus(4ba4b666-4513-46b8-b0f8-f39f56fb558f)\"" pod="openshift-multus/multus-zbfxr" podUID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.419662 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovnkube-controller/3.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.423797 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovn-acl-logging/0.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.424584 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6j5lq_2d662e59-55e9-45fa-870d-ad8c51a7a0ef/ovn-controller/0.log" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425654 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425702 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425732 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425743 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425649 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425868 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425931 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425756 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.425949 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426017 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426076 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" exitCode=0 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426096 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" exitCode=143 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426115 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" exitCode=143 Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426162 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426220 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426253 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426272 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426286 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426300 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426312 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426324 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426336 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426349 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426377 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426432 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426452 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426469 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426482 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426495 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426506 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426518 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426530 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426541 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426553 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426566 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426604 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426621 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426636 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426650 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426665 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426679 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426694 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426708 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426724 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426738 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426760 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6j5lq" event={"ID":"2d662e59-55e9-45fa-870d-ad8c51a7a0ef","Type":"ContainerDied","Data":"f48f416a055c9523d571a98976942868f588e6d93f4c1858e9aede4a05dc8947"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426788 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426806 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426888 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426906 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426921 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426936 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426951 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426984 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.426996 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.427008 4888 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.438253 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.509998 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6j5lq"] Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.513463 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6j5lq"] Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.514503 4888 scope.go:117] "RemoveContainer" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.573174 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.602135 4888 scope.go:117] "RemoveContainer" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.625118 4888 scope.go:117] "RemoveContainer" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.656905 4888 scope.go:117] "RemoveContainer" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.684298 4888 scope.go:117] "RemoveContainer" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.708949 4888 scope.go:117] "RemoveContainer" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.753965 4888 scope.go:117] "RemoveContainer" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.802003 4888 scope.go:117] "RemoveContainer" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.839659 4888 scope.go:117] "RemoveContainer" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.881885 4888 scope.go:117] "RemoveContainer" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.882461 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": container with ID starting with f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0 not found: ID does not exist" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.882496 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} err="failed to get container status \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": rpc error: code = NotFound desc = could not find container \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": container with ID starting with f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.882523 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.885836 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": container with ID starting with baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece not found: ID does not exist" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.885862 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} err="failed to get container status \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": rpc error: code = NotFound desc = could not find container \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": container with ID starting with baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.885891 4888 scope.go:117] "RemoveContainer" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.886108 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": container with ID starting with bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b not found: ID does not exist" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.886136 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} err="failed to get container status \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": rpc error: code = NotFound desc = could not find container \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": container with ID starting with bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.886149 4888 scope.go:117] "RemoveContainer" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.890038 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": container with ID starting with 20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74 not found: ID does not exist" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.890068 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} err="failed to get container status \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": rpc error: code = NotFound desc = could not find container \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": container with ID starting with 20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.890087 4888 scope.go:117] "RemoveContainer" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.894102 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": container with ID starting with 5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a not found: ID does not exist" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.894134 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} err="failed to get container status \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": rpc error: code = NotFound desc = could not find container \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": container with ID starting with 5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.894154 4888 scope.go:117] "RemoveContainer" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.894542 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": container with ID starting with e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a not found: ID does not exist" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.894595 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} err="failed to get container status \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": rpc error: code = NotFound desc = could not find container \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": container with ID starting with e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.894633 4888 scope.go:117] "RemoveContainer" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.895058 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": container with ID starting with eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e not found: ID does not exist" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895110 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} err="failed to get container status \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": rpc error: code = NotFound desc = could not find container \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": container with ID starting with eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895145 4888 scope.go:117] "RemoveContainer" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.895485 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": container with ID starting with 9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b not found: ID does not exist" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895512 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} err="failed to get container status \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": rpc error: code = NotFound desc = could not find container \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": container with ID starting with 9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895529 4888 scope.go:117] "RemoveContainer" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.895775 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": container with ID starting with e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c not found: ID does not exist" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895799 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} err="failed to get container status \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": rpc error: code = NotFound desc = could not find container \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": container with ID starting with e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.895828 4888 scope.go:117] "RemoveContainer" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: E1124 00:34:22.899307 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": container with ID starting with 06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48 not found: ID does not exist" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.899343 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} err="failed to get container status \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": rpc error: code = NotFound desc = could not find container \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": container with ID starting with 06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.899366 4888 scope.go:117] "RemoveContainer" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.902125 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} err="failed to get container status \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": rpc error: code = NotFound desc = could not find container \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": container with ID starting with f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.902156 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.905200 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} err="failed to get container status \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": rpc error: code = NotFound desc = could not find container \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": container with ID starting with baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.905253 4888 scope.go:117] "RemoveContainer" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.905643 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} err="failed to get container status \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": rpc error: code = NotFound desc = could not find container \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": container with ID starting with bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.905691 4888 scope.go:117] "RemoveContainer" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.906036 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} err="failed to get container status \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": rpc error: code = NotFound desc = could not find container \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": container with ID starting with 20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.906058 4888 scope.go:117] "RemoveContainer" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.908652 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} err="failed to get container status \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": rpc error: code = NotFound desc = could not find container \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": container with ID starting with 5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.908685 4888 scope.go:117] "RemoveContainer" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.909087 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} err="failed to get container status \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": rpc error: code = NotFound desc = could not find container \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": container with ID starting with e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.909126 4888 scope.go:117] "RemoveContainer" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.909395 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} err="failed to get container status \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": rpc error: code = NotFound desc = could not find container \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": container with ID starting with eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.909418 4888 scope.go:117] "RemoveContainer" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.912994 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} err="failed to get container status \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": rpc error: code = NotFound desc = could not find container \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": container with ID starting with 9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.913031 4888 scope.go:117] "RemoveContainer" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.915185 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} err="failed to get container status \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": rpc error: code = NotFound desc = could not find container \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": container with ID starting with e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.915234 4888 scope.go:117] "RemoveContainer" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.915641 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} err="failed to get container status \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": rpc error: code = NotFound desc = could not find container \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": container with ID starting with 06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.915690 4888 scope.go:117] "RemoveContainer" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.916413 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} err="failed to get container status \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": rpc error: code = NotFound desc = could not find container \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": container with ID starting with f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.916453 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.920262 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} err="failed to get container status \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": rpc error: code = NotFound desc = could not find container \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": container with ID starting with baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.920309 4888 scope.go:117] "RemoveContainer" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.920653 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} err="failed to get container status \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": rpc error: code = NotFound desc = could not find container \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": container with ID starting with bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.920686 4888 scope.go:117] "RemoveContainer" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.921130 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} err="failed to get container status \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": rpc error: code = NotFound desc = could not find container \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": container with ID starting with 20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.921176 4888 scope.go:117] "RemoveContainer" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.921690 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} err="failed to get container status \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": rpc error: code = NotFound desc = could not find container \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": container with ID starting with 5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.921714 4888 scope.go:117] "RemoveContainer" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.922046 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} err="failed to get container status \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": rpc error: code = NotFound desc = could not find container \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": container with ID starting with e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.922083 4888 scope.go:117] "RemoveContainer" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.922662 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} err="failed to get container status \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": rpc error: code = NotFound desc = could not find container \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": container with ID starting with eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.922684 4888 scope.go:117] "RemoveContainer" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923060 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} err="failed to get container status \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": rpc error: code = NotFound desc = could not find container \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": container with ID starting with 9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923088 4888 scope.go:117] "RemoveContainer" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923477 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} err="failed to get container status \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": rpc error: code = NotFound desc = could not find container \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": container with ID starting with e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923505 4888 scope.go:117] "RemoveContainer" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923921 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} err="failed to get container status \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": rpc error: code = NotFound desc = could not find container \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": container with ID starting with 06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.923943 4888 scope.go:117] "RemoveContainer" containerID="f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.924226 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0"} err="failed to get container status \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": rpc error: code = NotFound desc = could not find container \"f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0\": container with ID starting with f1a16d36814926caa5362554ace85d83ab2a32af83d5f044d26d8c83fe8ebbe0 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.924252 4888 scope.go:117] "RemoveContainer" containerID="baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928146 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece"} err="failed to get container status \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": rpc error: code = NotFound desc = could not find container \"baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece\": container with ID starting with baaab7b975579a0b4484fb5eb2fb629c1430955275cf64c717759f62fa1d4ece not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928197 4888 scope.go:117] "RemoveContainer" containerID="bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928523 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b"} err="failed to get container status \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": rpc error: code = NotFound desc = could not find container \"bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b\": container with ID starting with bf446d89f422591174953c7bca17db0e6bf6f746814d1b0fc555c2c4fde7137b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928548 4888 scope.go:117] "RemoveContainer" containerID="20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928847 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74"} err="failed to get container status \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": rpc error: code = NotFound desc = could not find container \"20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74\": container with ID starting with 20eb266ab4c5b46472d82e78c101e9cf517c34d9f98e55eebf9ad80e39743a74 not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.928903 4888 scope.go:117] "RemoveContainer" containerID="5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.929231 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a"} err="failed to get container status \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": rpc error: code = NotFound desc = could not find container \"5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a\": container with ID starting with 5dc44f440e6b50fc118c6ce87458162a678d50c46ae6b217aee2f5c68392de7a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.929254 4888 scope.go:117] "RemoveContainer" containerID="e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.933223 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a"} err="failed to get container status \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": rpc error: code = NotFound desc = could not find container \"e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a\": container with ID starting with e1f27781b19275c7c9189af0bb61ec02e2676fe6536a30b2a8d30cdb12615c6a not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.933262 4888 scope.go:117] "RemoveContainer" containerID="eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.933905 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e"} err="failed to get container status \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": rpc error: code = NotFound desc = could not find container \"eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e\": container with ID starting with eefc85294bb5dd8122d3e53495296191bb00aac5dca180adf9469a0cc273587e not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.933932 4888 scope.go:117] "RemoveContainer" containerID="9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.937126 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b"} err="failed to get container status \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": rpc error: code = NotFound desc = could not find container \"9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b\": container with ID starting with 9fa604c849bdfcf6ec82bc555ef39f9289edf6ebf007e5cdb22517a03519165b not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.937169 4888 scope.go:117] "RemoveContainer" containerID="e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.937592 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c"} err="failed to get container status \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": rpc error: code = NotFound desc = could not find container \"e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c\": container with ID starting with e48a9e0d784ef264c4d1723cbaa44ffa5ec8955459309fdbc9d073951364c65c not found: ID does not exist" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.937622 4888 scope.go:117] "RemoveContainer" containerID="06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48" Nov 24 00:34:22 crc kubenswrapper[4888]: I1124 00:34:22.937979 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48"} err="failed to get container status \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": rpc error: code = NotFound desc = could not find container \"06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48\": container with ID starting with 06acf034e931687010f7065d833e4c6a0c567712d342115fb6c0798e57d2ad48 not found: ID does not exist" Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.437102 4888 generic.go:334] "Generic (PLEG): container finished" podID="0f1d570f-0b3b-4553-a9df-a223f9c4c109" containerID="1c743b1b4bd282baa79ffdf03fd6aac895b8dc8df6aac7b8feb04b38c1a7f050" exitCode=0 Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.437179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerDied","Data":"1c743b1b4bd282baa79ffdf03fd6aac895b8dc8df6aac7b8feb04b38c1a7f050"} Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.437624 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"d93b321d2bb568ecbbf8d16c923093ee5d9442421e2ae7bf4f322faef3413ea0"} Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.439594 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/2.log" Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.592065 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.592118 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.592164 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.592900 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:34:23 crc kubenswrapper[4888]: I1124 00:34:23.592962 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde" gracePeriod=600 Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.253193 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d662e59-55e9-45fa-870d-ad8c51a7a0ef" path="/var/lib/kubelet/pods/2d662e59-55e9-45fa-870d-ad8c51a7a0ef/volumes" Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.448917 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"8eaa63200af2589799b07f9734ff99903bca889072327e28cb3636a37d9a4c63"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.448976 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"06ab56c0b9a97d588372a5e6bddc50b01f7b3dec6d3d94240bf2c0c09486fa47"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.448988 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"c431b90a3406ea4675902d1c59c76f677203ccd03019e3e5dc84dd0da03b31a8"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.448997 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"1a20e8496bdcec1679f8a3f388957c4a5c6d7e2cdd2edf46d0d3e280da8f0f2a"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.449008 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"e705ed2bcb4e24cfdb5f957e9305e17ed1ae59b96aa1bd3b43a6e40c24f7cbfa"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.449016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"bcfbd135af74e6146040dde540af5ba685d32d652226df1877e73350d4663f94"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.452748 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde" exitCode=0 Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.452803 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.452858 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4"} Nov 24 00:34:24 crc kubenswrapper[4888]: I1124 00:34:24.452883 4888 scope.go:117] "RemoveContainer" containerID="4aab818491112e354d6d855b3e2b8908ca4e7a27cbb56a222c2ba869d0f756a4" Nov 24 00:34:27 crc kubenswrapper[4888]: I1124 00:34:27.475882 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"2740679e1179dc2df1a1996a69b83cfc86baee9f7df54bca4c9ccd7a2c031afa"} Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.636873 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x"] Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.638061 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.640349 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-bslgg" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.640742 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.642720 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.692433 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf7pj\" (UniqueName: \"kubernetes.io/projected/85dc9246-ac2d-44f7-bda0-bdd6faa6a391-kube-api-access-kf7pj\") pod \"obo-prometheus-operator-668cf9dfbb-x756x\" (UID: \"85dc9246-ac2d-44f7-bda0-bdd6faa6a391\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.793775 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf7pj\" (UniqueName: \"kubernetes.io/projected/85dc9246-ac2d-44f7-bda0-bdd6faa6a391-kube-api-access-kf7pj\") pod \"obo-prometheus-operator-668cf9dfbb-x756x\" (UID: \"85dc9246-ac2d-44f7-bda0-bdd6faa6a391\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.802394 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2"] Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.803449 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.810786 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-gh9bm" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.814272 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd"] Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.815212 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.816169 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.822753 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf7pj\" (UniqueName: \"kubernetes.io/projected/85dc9246-ac2d-44f7-bda0-bdd6faa6a391-kube-api-access-kf7pj\") pod \"obo-prometheus-operator-668cf9dfbb-x756x\" (UID: \"85dc9246-ac2d-44f7-bda0-bdd6faa6a391\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.894657 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.895175 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.895226 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.895247 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.956698 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: E1124 00:34:28.981207 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(da0fb281323e444c4e1140d6bf79459540fe81503b722c87438f45614f856c1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:28 crc kubenswrapper[4888]: E1124 00:34:28.981296 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(da0fb281323e444c4e1140d6bf79459540fe81503b722c87438f45614f856c1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: E1124 00:34:28.981327 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(da0fb281323e444c4e1140d6bf79459540fe81503b722c87438f45614f856c1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:28 crc kubenswrapper[4888]: E1124 00:34:28.981374 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(da0fb281323e444c4e1140d6bf79459540fe81503b722c87438f45614f856c1b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" podUID="85dc9246-ac2d-44f7-bda0-bdd6faa6a391" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.996254 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.996310 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.996356 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.996381 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:28 crc kubenswrapper[4888]: I1124 00:34:28.996997 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vj7xs"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:28.998027 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.001469 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.001522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a11b0ce1-5305-4b79-a302-4839c19791e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2\" (UID: \"a11b0ce1-5305-4b79-a302-4839c19791e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.002437 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.006328 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd\" (UID: \"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.006334 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-xshlg" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.006441 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.097472 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9cc7d0-818c-484c-96fd-c22c437348bc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.097558 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tclf4\" (UniqueName: \"kubernetes.io/projected/2c9cc7d0-818c-484c-96fd-c22c437348bc-kube-api-access-tclf4\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.118004 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.140466 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(745181106b0b07a6c25bdbfaee1298c035af620fd184e38dca66c4c9d981ba13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.140589 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(745181106b0b07a6c25bdbfaee1298c035af620fd184e38dca66c4c9d981ba13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.140626 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(745181106b0b07a6c25bdbfaee1298c035af620fd184e38dca66c4c9d981ba13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.140704 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(745181106b0b07a6c25bdbfaee1298c035af620fd184e38dca66c4c9d981ba13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" podUID="a11b0ce1-5305-4b79-a302-4839c19791e0" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.160262 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.183975 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(0bef86cd8d2e75d56e51ae048a7b26c674f2ab5826d9c2f13571a343cfc1b3d6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.184062 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(0bef86cd8d2e75d56e51ae048a7b26c674f2ab5826d9c2f13571a343cfc1b3d6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.184088 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(0bef86cd8d2e75d56e51ae048a7b26c674f2ab5826d9c2f13571a343cfc1b3d6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.184163 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(0bef86cd8d2e75d56e51ae048a7b26c674f2ab5826d9c2f13571a343cfc1b3d6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" podUID="9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.192644 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-7j7tc"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.193719 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.198986 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9cc7d0-818c-484c-96fd-c22c437348bc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.199054 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tclf4\" (UniqueName: \"kubernetes.io/projected/2c9cc7d0-818c-484c-96fd-c22c437348bc-kube-api-access-tclf4\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.203389 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-65qvl" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.203700 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9cc7d0-818c-484c-96fd-c22c437348bc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.224411 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tclf4\" (UniqueName: \"kubernetes.io/projected/2c9cc7d0-818c-484c-96fd-c22c437348bc-kube-api-access-tclf4\") pod \"observability-operator-d8bb48f5d-vj7xs\" (UID: \"2c9cc7d0-818c-484c-96fd-c22c437348bc\") " pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.300001 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/920040ee-4308-431b-9f6d-eb16563acb3f-openshift-service-ca\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.300141 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldh49\" (UniqueName: \"kubernetes.io/projected/920040ee-4308-431b-9f6d-eb16563acb3f-kube-api-access-ldh49\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.334881 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.360418 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(e44ebfe92f06a8bbcf244d429e3cd458606bd93d543fcbaea64264f46904875b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.360520 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(e44ebfe92f06a8bbcf244d429e3cd458606bd93d543fcbaea64264f46904875b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.360551 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(e44ebfe92f06a8bbcf244d429e3cd458606bd93d543fcbaea64264f46904875b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.360616 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(e44ebfe92f06a8bbcf244d429e3cd458606bd93d543fcbaea64264f46904875b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" podUID="2c9cc7d0-818c-484c-96fd-c22c437348bc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.402056 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldh49\" (UniqueName: \"kubernetes.io/projected/920040ee-4308-431b-9f6d-eb16563acb3f-kube-api-access-ldh49\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.402219 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/920040ee-4308-431b-9f6d-eb16563acb3f-openshift-service-ca\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.403498 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/920040ee-4308-431b-9f6d-eb16563acb3f-openshift-service-ca\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.422687 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldh49\" (UniqueName: \"kubernetes.io/projected/920040ee-4308-431b-9f6d-eb16563acb3f-kube-api-access-ldh49\") pod \"perses-operator-5446b9c989-7j7tc\" (UID: \"920040ee-4308-431b-9f6d-eb16563acb3f\") " pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.491665 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" event={"ID":"0f1d570f-0b3b-4553-a9df-a223f9c4c109","Type":"ContainerStarted","Data":"fc09809c6ea2e01611ba6254342a31b359402ce754a23baf942575790a07d0d5"} Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.492049 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.492110 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.513726 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.529892 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.530581 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" podStartSLOduration=7.53055622 podStartE2EDuration="7.53055622s" podCreationTimestamp="2025-11-24 00:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:34:29.525983756 +0000 UTC m=+572.108667800" watchObservedRunningTime="2025-11-24 00:34:29.53055622 +0000 UTC m=+572.113240264" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.553784 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(506c7f23dc6a727ac39cf216ed90e179f0429da48ec04457614190d144570fc5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.554984 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(506c7f23dc6a727ac39cf216ed90e179f0429da48ec04457614190d144570fc5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.555042 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(506c7f23dc6a727ac39cf216ed90e179f0429da48ec04457614190d144570fc5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.555119 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(506c7f23dc6a727ac39cf216ed90e179f0429da48ec04457614190d144570fc5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" podUID="920040ee-4308-431b-9f6d-eb16563acb3f" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.877329 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.877989 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.878569 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.897820 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.897963 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.898451 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.899931 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(9ca7945f87f8b0ac1328c8bc43bfc9809d53268ad804fc67299c607baf8fb841): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.900034 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(9ca7945f87f8b0ac1328c8bc43bfc9809d53268ad804fc67299c607baf8fb841): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.900067 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(9ca7945f87f8b0ac1328c8bc43bfc9809d53268ad804fc67299c607baf8fb841): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.900139 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(9ca7945f87f8b0ac1328c8bc43bfc9809d53268ad804fc67299c607baf8fb841): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" podUID="a11b0ce1-5305-4b79-a302-4839c19791e0" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.914203 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.914362 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.914880 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.921873 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vj7xs"] Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.922034 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.922549 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: I1124 00:34:29.927233 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-7j7tc"] Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.946415 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(bdf9f79208107143dfc7d57ea3888e49a3d7a82481662e267591312fd99d7ea3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.946511 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(bdf9f79208107143dfc7d57ea3888e49a3d7a82481662e267591312fd99d7ea3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.946547 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(bdf9f79208107143dfc7d57ea3888e49a3d7a82481662e267591312fd99d7ea3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.946612 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(bdf9f79208107143dfc7d57ea3888e49a3d7a82481662e267591312fd99d7ea3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" podUID="85dc9246-ac2d-44f7-bda0-bdd6faa6a391" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.970056 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(ac5b846faadc930efdab6c086f280396d4e7553e71a785ad3f539344d3a1e00a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.970169 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(ac5b846faadc930efdab6c086f280396d4e7553e71a785ad3f539344d3a1e00a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.970203 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(ac5b846faadc930efdab6c086f280396d4e7553e71a785ad3f539344d3a1e00a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.970269 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(ac5b846faadc930efdab6c086f280396d4e7553e71a785ad3f539344d3a1e00a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" podUID="9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.985116 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(316a44b7637c029ff15b2c5865b27cf99cebfafe231126f1e6d3691322f0185b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.985209 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(316a44b7637c029ff15b2c5865b27cf99cebfafe231126f1e6d3691322f0185b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.985243 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(316a44b7637c029ff15b2c5865b27cf99cebfafe231126f1e6d3691322f0185b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:29 crc kubenswrapper[4888]: E1124 00:34:29.985309 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(316a44b7637c029ff15b2c5865b27cf99cebfafe231126f1e6d3691322f0185b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" podUID="2c9cc7d0-818c-484c-96fd-c22c437348bc" Nov 24 00:34:30 crc kubenswrapper[4888]: I1124 00:34:30.496923 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:30 crc kubenswrapper[4888]: I1124 00:34:30.497328 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:30 crc kubenswrapper[4888]: I1124 00:34:30.498878 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:30 crc kubenswrapper[4888]: I1124 00:34:30.524973 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:30 crc kubenswrapper[4888]: E1124 00:34:30.529175 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(64f9fb647eebc5389ff9def02167710973751daffdd7b3395a181888d37d262b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:30 crc kubenswrapper[4888]: E1124 00:34:30.529287 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(64f9fb647eebc5389ff9def02167710973751daffdd7b3395a181888d37d262b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:30 crc kubenswrapper[4888]: E1124 00:34:30.529321 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(64f9fb647eebc5389ff9def02167710973751daffdd7b3395a181888d37d262b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:30 crc kubenswrapper[4888]: E1124 00:34:30.529391 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(64f9fb647eebc5389ff9def02167710973751daffdd7b3395a181888d37d262b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" podUID="920040ee-4308-431b-9f6d-eb16563acb3f" Nov 24 00:34:36 crc kubenswrapper[4888]: I1124 00:34:36.245462 4888 scope.go:117] "RemoveContainer" containerID="2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799" Nov 24 00:34:36 crc kubenswrapper[4888]: E1124 00:34:36.247388 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-zbfxr_openshift-multus(4ba4b666-4513-46b8-b0f8-f39f56fb558f)\"" pod="openshift-multus/multus-zbfxr" podUID="4ba4b666-4513-46b8-b0f8-f39f56fb558f" Nov 24 00:34:41 crc kubenswrapper[4888]: I1124 00:34:41.244332 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:41 crc kubenswrapper[4888]: I1124 00:34:41.245580 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:41 crc kubenswrapper[4888]: E1124 00:34:41.287487 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(dd9c94447b626440a6950a4d8c42638509e4d4661334c9d9daf3633cd2851774): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:41 crc kubenswrapper[4888]: E1124 00:34:41.287592 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(dd9c94447b626440a6950a4d8c42638509e4d4661334c9d9daf3633cd2851774): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:41 crc kubenswrapper[4888]: E1124 00:34:41.287626 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(dd9c94447b626440a6950a4d8c42638509e4d4661334c9d9daf3633cd2851774): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:41 crc kubenswrapper[4888]: E1124 00:34:41.287693 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-vj7xs_openshift-operators(2c9cc7d0-818c-484c-96fd-c22c437348bc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-vj7xs_openshift-operators_2c9cc7d0-818c-484c-96fd-c22c437348bc_0(dd9c94447b626440a6950a4d8c42638509e4d4661334c9d9daf3633cd2851774): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" podUID="2c9cc7d0-818c-484c-96fd-c22c437348bc" Nov 24 00:34:42 crc kubenswrapper[4888]: I1124 00:34:42.245064 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:42 crc kubenswrapper[4888]: I1124 00:34:42.245884 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:42 crc kubenswrapper[4888]: E1124 00:34:42.275407 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(77364089fb6729c47b2fc9f1b9d876821d37d717f634f8e04362ad9c58147cea): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:42 crc kubenswrapper[4888]: E1124 00:34:42.275501 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(77364089fb6729c47b2fc9f1b9d876821d37d717f634f8e04362ad9c58147cea): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:42 crc kubenswrapper[4888]: E1124 00:34:42.275537 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(77364089fb6729c47b2fc9f1b9d876821d37d717f634f8e04362ad9c58147cea): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:42 crc kubenswrapper[4888]: E1124 00:34:42.275601 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-7j7tc_openshift-operators(920040ee-4308-431b-9f6d-eb16563acb3f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-7j7tc_openshift-operators_920040ee-4308-431b-9f6d-eb16563acb3f_0(77364089fb6729c47b2fc9f1b9d876821d37d717f634f8e04362ad9c58147cea): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" podUID="920040ee-4308-431b-9f6d-eb16563acb3f" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.244692 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.244865 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.245241 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.245391 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.245534 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:43 crc kubenswrapper[4888]: I1124 00:34:43.245843 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.302212 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(6199fbf4c657a558769142298cf6b87c516e19011be15d7228f95578e6ca0333): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.302307 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(6199fbf4c657a558769142298cf6b87c516e19011be15d7228f95578e6ca0333): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.302367 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(6199fbf4c657a558769142298cf6b87c516e19011be15d7228f95578e6ca0333): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.302426 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators(85dc9246-ac2d-44f7-bda0-bdd6faa6a391)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-x756x_openshift-operators_85dc9246-ac2d-44f7-bda0-bdd6faa6a391_0(6199fbf4c657a558769142298cf6b87c516e19011be15d7228f95578e6ca0333): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" podUID="85dc9246-ac2d-44f7-bda0-bdd6faa6a391" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.312554 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(721874cf4ea14982b1a6a4510e1e3e9d402cdfdee7ccb20946456b9b5065978a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.312641 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(721874cf4ea14982b1a6a4510e1e3e9d402cdfdee7ccb20946456b9b5065978a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.312666 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(721874cf4ea14982b1a6a4510e1e3e9d402cdfdee7ccb20946456b9b5065978a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.312734 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators(a11b0ce1-5305-4b79-a302-4839c19791e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_openshift-operators_a11b0ce1-5305-4b79-a302-4839c19791e0_0(721874cf4ea14982b1a6a4510e1e3e9d402cdfdee7ccb20946456b9b5065978a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" podUID="a11b0ce1-5305-4b79-a302-4839c19791e0" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.320724 4888 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(2c25be77548ddd16b123244a4f464264349c7438774a58cf9247ec463cd628cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.320793 4888 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(2c25be77548ddd16b123244a4f464264349c7438774a58cf9247ec463cd628cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.321622 4888 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(2c25be77548ddd16b123244a4f464264349c7438774a58cf9247ec463cd628cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:43 crc kubenswrapper[4888]: E1124 00:34:43.321700 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators(9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_openshift-operators_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4_0(2c25be77548ddd16b123244a4f464264349c7438774a58cf9247ec463cd628cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" podUID="9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4" Nov 24 00:34:50 crc kubenswrapper[4888]: I1124 00:34:50.245110 4888 scope.go:117] "RemoveContainer" containerID="2ad525da8c838fdcb2da39277ecb308f1db815daec74c1ea41918a470845b799" Nov 24 00:34:50 crc kubenswrapper[4888]: I1124 00:34:50.612941 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zbfxr_4ba4b666-4513-46b8-b0f8-f39f56fb558f/kube-multus/2.log" Nov 24 00:34:50 crc kubenswrapper[4888]: I1124 00:34:50.613012 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zbfxr" event={"ID":"4ba4b666-4513-46b8-b0f8-f39f56fb558f","Type":"ContainerStarted","Data":"187e0081019561be028a28312c3ff7384973a0b685f6c53eaa41f7fd32fe7b07"} Nov 24 00:34:52 crc kubenswrapper[4888]: I1124 00:34:52.471780 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6pmlv" Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.244239 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.245078 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.245743 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.246714 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.515324 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-7j7tc"] Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.565419 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd"] Nov 24 00:34:54 crc kubenswrapper[4888]: W1124 00:34:54.575495 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a7b0bb7_bf4c_47f3_a39d_b05a758f50f4.slice/crio-955db0362a27654aafa2655c49543b10e25f5f4ee79c1e14812aca0c6aae502f WatchSource:0}: Error finding container 955db0362a27654aafa2655c49543b10e25f5f4ee79c1e14812aca0c6aae502f: Status 404 returned error can't find the container with id 955db0362a27654aafa2655c49543b10e25f5f4ee79c1e14812aca0c6aae502f Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.635758 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" event={"ID":"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4","Type":"ContainerStarted","Data":"955db0362a27654aafa2655c49543b10e25f5f4ee79c1e14812aca0c6aae502f"} Nov 24 00:34:54 crc kubenswrapper[4888]: I1124 00:34:54.636918 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" event={"ID":"920040ee-4308-431b-9f6d-eb16563acb3f","Type":"ContainerStarted","Data":"0cb62545d1390e700444a7da3ca2f67cb552d37454f91ddc2a1a8fec78680ac0"} Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.245217 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.245333 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.245961 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.246063 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.511979 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vj7xs"] Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.557904 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x"] Nov 24 00:34:55 crc kubenswrapper[4888]: W1124 00:34:55.564860 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85dc9246_ac2d_44f7_bda0_bdd6faa6a391.slice/crio-63c38bbb0b6e6cc9283dfc8e4a9c02541e9c287d21914aa521723f9ca5464ca5 WatchSource:0}: Error finding container 63c38bbb0b6e6cc9283dfc8e4a9c02541e9c287d21914aa521723f9ca5464ca5: Status 404 returned error can't find the container with id 63c38bbb0b6e6cc9283dfc8e4a9c02541e9c287d21914aa521723f9ca5464ca5 Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.645875 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" event={"ID":"85dc9246-ac2d-44f7-bda0-bdd6faa6a391","Type":"ContainerStarted","Data":"63c38bbb0b6e6cc9283dfc8e4a9c02541e9c287d21914aa521723f9ca5464ca5"} Nov 24 00:34:55 crc kubenswrapper[4888]: I1124 00:34:55.647774 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" event={"ID":"2c9cc7d0-818c-484c-96fd-c22c437348bc","Type":"ContainerStarted","Data":"9c519d5bdfbd048c11911d618f41c558cb5adf24cdcee930cf79cba106622055"} Nov 24 00:34:56 crc kubenswrapper[4888]: I1124 00:34:56.245241 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:56 crc kubenswrapper[4888]: I1124 00:34:56.246380 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" Nov 24 00:34:56 crc kubenswrapper[4888]: I1124 00:34:56.694645 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2"] Nov 24 00:34:56 crc kubenswrapper[4888]: W1124 00:34:56.709447 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda11b0ce1_5305_4b79_a302_4839c19791e0.slice/crio-c62b9b73ab4ea210999915875fff0d5d9d3d047aee2d27e290651b1d96ab5612 WatchSource:0}: Error finding container c62b9b73ab4ea210999915875fff0d5d9d3d047aee2d27e290651b1d96ab5612: Status 404 returned error can't find the container with id c62b9b73ab4ea210999915875fff0d5d9d3d047aee2d27e290651b1d96ab5612 Nov 24 00:34:57 crc kubenswrapper[4888]: I1124 00:34:57.680063 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" event={"ID":"a11b0ce1-5305-4b79-a302-4839c19791e0","Type":"ContainerStarted","Data":"c62b9b73ab4ea210999915875fff0d5d9d3d047aee2d27e290651b1d96ab5612"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.735467 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" event={"ID":"920040ee-4308-431b-9f6d-eb16563acb3f","Type":"ContainerStarted","Data":"66f39c8eff49e7210f6090b4098188ef3a1d708e1ffd81c225cc947f3eac1bf0"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.735941 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.738760 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" event={"ID":"85dc9246-ac2d-44f7-bda0-bdd6faa6a391","Type":"ContainerStarted","Data":"d79af8a06fa5e42ca93c4ad8cad6d0b0a85ff07afec6cd448e6dd2c4ba3d7b81"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.740983 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" event={"ID":"a11b0ce1-5305-4b79-a302-4839c19791e0","Type":"ContainerStarted","Data":"ac9ea50164dde3d3638b799916460721a8f316a1e5e2ccf2441edd9b5bd10cb6"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.744038 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" event={"ID":"9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4","Type":"ContainerStarted","Data":"fbef358aef929acae17513e44d1e966e58e7cce09afbe613e9b03b56fe3ca06c"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.747011 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" event={"ID":"2c9cc7d0-818c-484c-96fd-c22c437348bc","Type":"ContainerStarted","Data":"9497bebf52caba49a86ba8a76422666fdf70f7d8b241ea149c640b79edb6fba1"} Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.747216 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.750173 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.779797 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" podStartSLOduration=26.17376183 podStartE2EDuration="36.779776793s" podCreationTimestamp="2025-11-24 00:34:29 +0000 UTC" firstStartedPulling="2025-11-24 00:34:54.528534026 +0000 UTC m=+597.111218080" lastFinishedPulling="2025-11-24 00:35:05.134548989 +0000 UTC m=+607.717233043" observedRunningTime="2025-11-24 00:35:05.77523846 +0000 UTC m=+608.357922544" watchObservedRunningTime="2025-11-24 00:35:05.779776793 +0000 UTC m=+608.362460847" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.805747 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-vj7xs" podStartSLOduration=28.197843161 podStartE2EDuration="37.805729511s" podCreationTimestamp="2025-11-24 00:34:28 +0000 UTC" firstStartedPulling="2025-11-24 00:34:55.526984036 +0000 UTC m=+598.109668080" lastFinishedPulling="2025-11-24 00:35:05.134870386 +0000 UTC m=+607.717554430" observedRunningTime="2025-11-24 00:35:05.803886159 +0000 UTC m=+608.386570204" watchObservedRunningTime="2025-11-24 00:35:05.805729511 +0000 UTC m=+608.388413545" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.884170 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd" podStartSLOduration=27.350364193 podStartE2EDuration="37.884143789s" podCreationTimestamp="2025-11-24 00:34:28 +0000 UTC" firstStartedPulling="2025-11-24 00:34:54.578914818 +0000 UTC m=+597.161598862" lastFinishedPulling="2025-11-24 00:35:05.112694414 +0000 UTC m=+607.695378458" observedRunningTime="2025-11-24 00:35:05.880036575 +0000 UTC m=+608.462720629" watchObservedRunningTime="2025-11-24 00:35:05.884143789 +0000 UTC m=+608.466827843" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.912090 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2" podStartSLOduration=29.506601752 podStartE2EDuration="37.912065521s" podCreationTimestamp="2025-11-24 00:34:28 +0000 UTC" firstStartedPulling="2025-11-24 00:34:56.712045584 +0000 UTC m=+599.294729648" lastFinishedPulling="2025-11-24 00:35:05.117509383 +0000 UTC m=+607.700193417" observedRunningTime="2025-11-24 00:35:05.909843111 +0000 UTC m=+608.492527155" watchObservedRunningTime="2025-11-24 00:35:05.912065521 +0000 UTC m=+608.494749565" Nov 24 00:35:05 crc kubenswrapper[4888]: I1124 00:35:05.950194 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-x756x" podStartSLOduration=28.434450073 podStartE2EDuration="37.950172045s" podCreationTimestamp="2025-11-24 00:34:28 +0000 UTC" firstStartedPulling="2025-11-24 00:34:55.568734612 +0000 UTC m=+598.151418666" lastFinishedPulling="2025-11-24 00:35:05.084456594 +0000 UTC m=+607.667140638" observedRunningTime="2025-11-24 00:35:05.946542413 +0000 UTC m=+608.529226447" watchObservedRunningTime="2025-11-24 00:35:05.950172045 +0000 UTC m=+608.532856079" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.577932 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vjmm8"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.579479 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.584404 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.584559 4888 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jgx7w" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.584704 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.601082 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2ls5n"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.607848 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2ls5n" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.618021 4888 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-slx5k" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.625824 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vjmm8"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.630735 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g8kdr"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.632561 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.636359 4888 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6qrjr" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.637580 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2ls5n"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.657309 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g8kdr"] Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.668226 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz9nm\" (UniqueName: \"kubernetes.io/projected/087d3537-623e-4f8b-be5b-b9342a2cf19f-kube-api-access-qz9nm\") pod \"cert-manager-5b446d88c5-2ls5n\" (UID: \"087d3537-623e-4f8b-be5b-b9342a2cf19f\") " pod="cert-manager/cert-manager-5b446d88c5-2ls5n" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.668343 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w54v\" (UniqueName: \"kubernetes.io/projected/6a61128e-aa42-4201-87ae-856cc4e6f682-kube-api-access-4w54v\") pod \"cert-manager-webhook-5655c58dd6-g8kdr\" (UID: \"6a61128e-aa42-4201-87ae-856cc4e6f682\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.668440 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmf7m\" (UniqueName: \"kubernetes.io/projected/25c1c650-be40-437b-9ea3-8e9a60b2f2f2-kube-api-access-vmf7m\") pod \"cert-manager-cainjector-7f985d654d-vjmm8\" (UID: \"25c1c650-be40-437b-9ea3-8e9a60b2f2f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.770985 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz9nm\" (UniqueName: \"kubernetes.io/projected/087d3537-623e-4f8b-be5b-b9342a2cf19f-kube-api-access-qz9nm\") pod \"cert-manager-5b446d88c5-2ls5n\" (UID: \"087d3537-623e-4f8b-be5b-b9342a2cf19f\") " pod="cert-manager/cert-manager-5b446d88c5-2ls5n" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.771088 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w54v\" (UniqueName: \"kubernetes.io/projected/6a61128e-aa42-4201-87ae-856cc4e6f682-kube-api-access-4w54v\") pod \"cert-manager-webhook-5655c58dd6-g8kdr\" (UID: \"6a61128e-aa42-4201-87ae-856cc4e6f682\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.771159 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmf7m\" (UniqueName: \"kubernetes.io/projected/25c1c650-be40-437b-9ea3-8e9a60b2f2f2-kube-api-access-vmf7m\") pod \"cert-manager-cainjector-7f985d654d-vjmm8\" (UID: \"25c1c650-be40-437b-9ea3-8e9a60b2f2f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.814035 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmf7m\" (UniqueName: \"kubernetes.io/projected/25c1c650-be40-437b-9ea3-8e9a60b2f2f2-kube-api-access-vmf7m\") pod \"cert-manager-cainjector-7f985d654d-vjmm8\" (UID: \"25c1c650-be40-437b-9ea3-8e9a60b2f2f2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.814455 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w54v\" (UniqueName: \"kubernetes.io/projected/6a61128e-aa42-4201-87ae-856cc4e6f682-kube-api-access-4w54v\") pod \"cert-manager-webhook-5655c58dd6-g8kdr\" (UID: \"6a61128e-aa42-4201-87ae-856cc4e6f682\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.824531 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz9nm\" (UniqueName: \"kubernetes.io/projected/087d3537-623e-4f8b-be5b-b9342a2cf19f-kube-api-access-qz9nm\") pod \"cert-manager-5b446d88c5-2ls5n\" (UID: \"087d3537-623e-4f8b-be5b-b9342a2cf19f\") " pod="cert-manager/cert-manager-5b446d88c5-2ls5n" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.919107 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.934995 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2ls5n" Nov 24 00:35:13 crc kubenswrapper[4888]: I1124 00:35:13.961454 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.179793 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vjmm8"] Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.225795 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2ls5n"] Nov 24 00:35:14 crc kubenswrapper[4888]: W1124 00:35:14.232394 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod087d3537_623e_4f8b_be5b_b9342a2cf19f.slice/crio-1734a4c8276ffd63ad76c571091296960a6bbf3f6adc204e95e92e6a42698d26 WatchSource:0}: Error finding container 1734a4c8276ffd63ad76c571091296960a6bbf3f6adc204e95e92e6a42698d26: Status 404 returned error can't find the container with id 1734a4c8276ffd63ad76c571091296960a6bbf3f6adc204e95e92e6a42698d26 Nov 24 00:35:14 crc kubenswrapper[4888]: W1124 00:35:14.262141 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a61128e_aa42_4201_87ae_856cc4e6f682.slice/crio-ed55b27a7ab95a59e2d17279549222f49ef83bce77cf6e4504250c53e32ef76b WatchSource:0}: Error finding container ed55b27a7ab95a59e2d17279549222f49ef83bce77cf6e4504250c53e32ef76b: Status 404 returned error can't find the container with id ed55b27a7ab95a59e2d17279549222f49ef83bce77cf6e4504250c53e32ef76b Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.263526 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-g8kdr"] Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.824186 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" event={"ID":"6a61128e-aa42-4201-87ae-856cc4e6f682","Type":"ContainerStarted","Data":"ed55b27a7ab95a59e2d17279549222f49ef83bce77cf6e4504250c53e32ef76b"} Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.826029 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2ls5n" event={"ID":"087d3537-623e-4f8b-be5b-b9342a2cf19f","Type":"ContainerStarted","Data":"1734a4c8276ffd63ad76c571091296960a6bbf3f6adc204e95e92e6a42698d26"} Nov 24 00:35:14 crc kubenswrapper[4888]: I1124 00:35:14.827250 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" event={"ID":"25c1c650-be40-437b-9ea3-8e9a60b2f2f2","Type":"ContainerStarted","Data":"0f8e03a3d0c3b1751a20c051ffd78ec6a3dab1c318ca3f79b67c1480431b2a81"} Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.879313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" event={"ID":"25c1c650-be40-437b-9ea3-8e9a60b2f2f2","Type":"ContainerStarted","Data":"e9be87b3e3991d420dd6907d1b3aafa9482be8ef1379d861568a3563798430aa"} Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.883268 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" event={"ID":"6a61128e-aa42-4201-87ae-856cc4e6f682","Type":"ContainerStarted","Data":"6c61cf7d89d0a95886936fdb181f835e9c3330b000da6c6bae7d7e667b9b1c33"} Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.883402 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.885777 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2ls5n" event={"ID":"087d3537-623e-4f8b-be5b-b9342a2cf19f","Type":"ContainerStarted","Data":"54d001ecbfe6bc1284bdc53f82fe1e32f33352b199814550be1720c4204eeaf5"} Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.904158 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-vjmm8" podStartSLOduration=2.2032817749999998 podStartE2EDuration="5.904128104s" podCreationTimestamp="2025-11-24 00:35:13 +0000 UTC" firstStartedPulling="2025-11-24 00:35:14.195925843 +0000 UTC m=+616.778609887" lastFinishedPulling="2025-11-24 00:35:17.896772172 +0000 UTC m=+620.479456216" observedRunningTime="2025-11-24 00:35:18.897579265 +0000 UTC m=+621.480263319" watchObservedRunningTime="2025-11-24 00:35:18.904128104 +0000 UTC m=+621.486812188" Nov 24 00:35:18 crc kubenswrapper[4888]: I1124 00:35:18.940627 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" podStartSLOduration=2.247254842 podStartE2EDuration="5.940601231s" podCreationTimestamp="2025-11-24 00:35:13 +0000 UTC" firstStartedPulling="2025-11-24 00:35:14.264764243 +0000 UTC m=+616.847448287" lastFinishedPulling="2025-11-24 00:35:17.958110632 +0000 UTC m=+620.540794676" observedRunningTime="2025-11-24 00:35:18.921278373 +0000 UTC m=+621.503962407" watchObservedRunningTime="2025-11-24 00:35:18.940601231 +0000 UTC m=+621.523285275" Nov 24 00:35:19 crc kubenswrapper[4888]: I1124 00:35:19.517148 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-7j7tc" Nov 24 00:35:19 crc kubenswrapper[4888]: I1124 00:35:19.542685 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-2ls5n" podStartSLOduration=2.7587362840000003 podStartE2EDuration="6.542652726s" podCreationTimestamp="2025-11-24 00:35:13 +0000 UTC" firstStartedPulling="2025-11-24 00:35:14.23460382 +0000 UTC m=+616.817287864" lastFinishedPulling="2025-11-24 00:35:18.018520262 +0000 UTC m=+620.601204306" observedRunningTime="2025-11-24 00:35:18.943484846 +0000 UTC m=+621.526168890" watchObservedRunningTime="2025-11-24 00:35:19.542652726 +0000 UTC m=+622.125336810" Nov 24 00:35:23 crc kubenswrapper[4888]: I1124 00:35:23.965008 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-g8kdr" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.425210 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss"] Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.427796 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.430198 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.436495 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss"] Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.567287 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhbh2\" (UniqueName: \"kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.567396 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.567435 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.576687 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm"] Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.578009 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.589362 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm"] Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.669237 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhbh2\" (UniqueName: \"kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.669331 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.669358 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.670159 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.670306 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.692752 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhbh2\" (UniqueName: \"kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.747947 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.770559 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.770625 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz9lp\" (UniqueName: \"kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.770985 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.874705 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.875188 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz9lp\" (UniqueName: \"kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.875244 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.875639 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.875710 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:50 crc kubenswrapper[4888]: I1124 00:35:50.910077 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz9lp\" (UniqueName: \"kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:51 crc kubenswrapper[4888]: I1124 00:35:51.070919 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss"] Nov 24 00:35:51 crc kubenswrapper[4888]: I1124 00:35:51.144585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" event={"ID":"25b7bdca-37de-4056-8f0f-781a398ad19e","Type":"ContainerStarted","Data":"8a2bbb5922e50f7b6b8b98ad2eadcd461b0ed1dd46d1702562a3fe27439c82b0"} Nov 24 00:35:51 crc kubenswrapper[4888]: I1124 00:35:51.194160 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:51 crc kubenswrapper[4888]: I1124 00:35:51.479774 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm"] Nov 24 00:35:51 crc kubenswrapper[4888]: W1124 00:35:51.492129 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff873a77_87b3_42db_8bc1_acdcca37ef42.slice/crio-41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead WatchSource:0}: Error finding container 41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead: Status 404 returned error can't find the container with id 41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead Nov 24 00:35:52 crc kubenswrapper[4888]: I1124 00:35:52.154957 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerID="c1de27d2fb2364c6cc4423e5a7bce48cd945562a34501743992a016e792c77fe" exitCode=0 Nov 24 00:35:52 crc kubenswrapper[4888]: I1124 00:35:52.155058 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" event={"ID":"ff873a77-87b3-42db-8bc1-acdcca37ef42","Type":"ContainerDied","Data":"c1de27d2fb2364c6cc4423e5a7bce48cd945562a34501743992a016e792c77fe"} Nov 24 00:35:52 crc kubenswrapper[4888]: I1124 00:35:52.155105 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" event={"ID":"ff873a77-87b3-42db-8bc1-acdcca37ef42","Type":"ContainerStarted","Data":"41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead"} Nov 24 00:35:52 crc kubenswrapper[4888]: I1124 00:35:52.162275 4888 generic.go:334] "Generic (PLEG): container finished" podID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerID="880f25a315545d226a5c747ad3de748f9972ec73296bd571f5aeabcc5ae3861d" exitCode=0 Nov 24 00:35:52 crc kubenswrapper[4888]: I1124 00:35:52.162414 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" event={"ID":"25b7bdca-37de-4056-8f0f-781a398ad19e","Type":"ContainerDied","Data":"880f25a315545d226a5c747ad3de748f9972ec73296bd571f5aeabcc5ae3861d"} Nov 24 00:35:54 crc kubenswrapper[4888]: I1124 00:35:54.182125 4888 generic.go:334] "Generic (PLEG): container finished" podID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerID="186265093e600a96ab1206f66841cf506dbae6139dfe70fec39e10184cb25c80" exitCode=0 Nov 24 00:35:54 crc kubenswrapper[4888]: I1124 00:35:54.182242 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" event={"ID":"25b7bdca-37de-4056-8f0f-781a398ad19e","Type":"ContainerDied","Data":"186265093e600a96ab1206f66841cf506dbae6139dfe70fec39e10184cb25c80"} Nov 24 00:35:54 crc kubenswrapper[4888]: I1124 00:35:54.185612 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerID="c297cc6d855346247619efa74f354dcda0c36e1f9c51c505dbe68118c46013d9" exitCode=0 Nov 24 00:35:54 crc kubenswrapper[4888]: I1124 00:35:54.185687 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" event={"ID":"ff873a77-87b3-42db-8bc1-acdcca37ef42","Type":"ContainerDied","Data":"c297cc6d855346247619efa74f354dcda0c36e1f9c51c505dbe68118c46013d9"} Nov 24 00:35:55 crc kubenswrapper[4888]: I1124 00:35:55.198262 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerID="595d5329a5192cc55a834757ef1ce09e30923e31db79d04f71aaed6b4bc66a10" exitCode=0 Nov 24 00:35:55 crc kubenswrapper[4888]: I1124 00:35:55.198402 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" event={"ID":"ff873a77-87b3-42db-8bc1-acdcca37ef42","Type":"ContainerDied","Data":"595d5329a5192cc55a834757ef1ce09e30923e31db79d04f71aaed6b4bc66a10"} Nov 24 00:35:55 crc kubenswrapper[4888]: I1124 00:35:55.202380 4888 generic.go:334] "Generic (PLEG): container finished" podID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerID="98b2f6d832054ff24f49cf03bcf6fd91bced35f9fd42219ba74b23fcd9141d3f" exitCode=0 Nov 24 00:35:55 crc kubenswrapper[4888]: I1124 00:35:55.202459 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" event={"ID":"25b7bdca-37de-4056-8f0f-781a398ad19e","Type":"ContainerDied","Data":"98b2f6d832054ff24f49cf03bcf6fd91bced35f9fd42219ba74b23fcd9141d3f"} Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.551222 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.630293 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.671615 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle\") pod \"ff873a77-87b3-42db-8bc1-acdcca37ef42\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.671686 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util\") pod \"ff873a77-87b3-42db-8bc1-acdcca37ef42\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.671725 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9lp\" (UniqueName: \"kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp\") pod \"ff873a77-87b3-42db-8bc1-acdcca37ef42\" (UID: \"ff873a77-87b3-42db-8bc1-acdcca37ef42\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.674236 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle" (OuterVolumeSpecName: "bundle") pod "ff873a77-87b3-42db-8bc1-acdcca37ef42" (UID: "ff873a77-87b3-42db-8bc1-acdcca37ef42"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.682037 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp" (OuterVolumeSpecName: "kube-api-access-lz9lp") pod "ff873a77-87b3-42db-8bc1-acdcca37ef42" (UID: "ff873a77-87b3-42db-8bc1-acdcca37ef42"). InnerVolumeSpecName "kube-api-access-lz9lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.687413 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util" (OuterVolumeSpecName: "util") pod "ff873a77-87b3-42db-8bc1-acdcca37ef42" (UID: "ff873a77-87b3-42db-8bc1-acdcca37ef42"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.772957 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhbh2\" (UniqueName: \"kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2\") pod \"25b7bdca-37de-4056-8f0f-781a398ad19e\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.773042 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle\") pod \"25b7bdca-37de-4056-8f0f-781a398ad19e\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.773110 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util\") pod \"25b7bdca-37de-4056-8f0f-781a398ad19e\" (UID: \"25b7bdca-37de-4056-8f0f-781a398ad19e\") " Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.773418 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.773438 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff873a77-87b3-42db-8bc1-acdcca37ef42-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.773449 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9lp\" (UniqueName: \"kubernetes.io/projected/ff873a77-87b3-42db-8bc1-acdcca37ef42-kube-api-access-lz9lp\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.774135 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle" (OuterVolumeSpecName: "bundle") pod "25b7bdca-37de-4056-8f0f-781a398ad19e" (UID: "25b7bdca-37de-4056-8f0f-781a398ad19e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.777217 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2" (OuterVolumeSpecName: "kube-api-access-xhbh2") pod "25b7bdca-37de-4056-8f0f-781a398ad19e" (UID: "25b7bdca-37de-4056-8f0f-781a398ad19e"). InnerVolumeSpecName "kube-api-access-xhbh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.875088 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhbh2\" (UniqueName: \"kubernetes.io/projected/25b7bdca-37de-4056-8f0f-781a398ad19e-kube-api-access-xhbh2\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:56 crc kubenswrapper[4888]: I1124 00:35:56.875120 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.080438 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util" (OuterVolumeSpecName: "util") pod "25b7bdca-37de-4056-8f0f-781a398ad19e" (UID: "25b7bdca-37de-4056-8f0f-781a398ad19e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.180815 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25b7bdca-37de-4056-8f0f-781a398ad19e-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.224292 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.224184 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm" event={"ID":"ff873a77-87b3-42db-8bc1-acdcca37ef42","Type":"ContainerDied","Data":"41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead"} Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.224794 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41cfa9ea6516f64ee28efe979ccb70e885c1afc271690e3ac5d687c0d41f1ead" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.227949 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" event={"ID":"25b7bdca-37de-4056-8f0f-781a398ad19e","Type":"ContainerDied","Data":"8a2bbb5922e50f7b6b8b98ad2eadcd461b0ed1dd46d1702562a3fe27439c82b0"} Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.228064 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss" Nov 24 00:35:57 crc kubenswrapper[4888]: I1124 00:35:57.228077 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2bbb5922e50f7b6b8b98ad2eadcd461b0ed1dd46d1702562a3fe27439c82b0" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.735326 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc"] Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736361 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="util" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736382 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="util" Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736410 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="pull" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736417 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="pull" Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736429 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736438 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736454 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736462 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736473 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="util" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736481 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="util" Nov 24 00:36:07 crc kubenswrapper[4888]: E1124 00:36:07.736493 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="pull" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736500 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="pull" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736665 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b7bdca-37de-4056-8f0f-781a398ad19e" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.736681 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff873a77-87b3-42db-8bc1-acdcca37ef42" containerName="extract" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.737568 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.740326 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-tvfgj" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.740443 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.740526 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.741319 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.741898 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.742291 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.758458 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc"] Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.788293 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-webhook-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.788415 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9m25\" (UniqueName: \"kubernetes.io/projected/d9fc74ad-5691-4d31-a9b8-9bab98d49121-kube-api-access-c9m25\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.788525 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.788555 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d9fc74ad-5691-4d31-a9b8-9bab98d49121-manager-config\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.788724 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-apiservice-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.889921 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-apiservice-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.889980 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9m25\" (UniqueName: \"kubernetes.io/projected/d9fc74ad-5691-4d31-a9b8-9bab98d49121-kube-api-access-c9m25\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.890400 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-webhook-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.891215 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.891253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d9fc74ad-5691-4d31-a9b8-9bab98d49121-manager-config\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.892113 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d9fc74ad-5691-4d31-a9b8-9bab98d49121-manager-config\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.897046 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-webhook-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.898089 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.905509 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9fc74ad-5691-4d31-a9b8-9bab98d49121-apiservice-cert\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:07 crc kubenswrapper[4888]: I1124 00:36:07.907507 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9m25\" (UniqueName: \"kubernetes.io/projected/d9fc74ad-5691-4d31-a9b8-9bab98d49121-kube-api-access-c9m25\") pod \"loki-operator-controller-manager-648dc49c56-gnzhc\" (UID: \"d9fc74ad-5691-4d31-a9b8-9bab98d49121\") " pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:08 crc kubenswrapper[4888]: I1124 00:36:08.053889 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:08 crc kubenswrapper[4888]: I1124 00:36:08.627868 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc"] Nov 24 00:36:09 crc kubenswrapper[4888]: I1124 00:36:09.326449 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" event={"ID":"d9fc74ad-5691-4d31-a9b8-9bab98d49121","Type":"ContainerStarted","Data":"44598d2ce363e14e1650d52a2f612aa6466089e0198a2e82ce5a6cf7e79cbf41"} Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.729958 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lv57g"] Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.731191 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.733209 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-vhs7w" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.733525 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.733558 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.745474 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lv57g"] Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.839164 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfhnq\" (UniqueName: \"kubernetes.io/projected/2b99942b-b619-4a73-bae0-b3cd34990e42-kube-api-access-kfhnq\") pod \"cluster-logging-operator-ff9846bd-lv57g\" (UID: \"2b99942b-b619-4a73-bae0-b3cd34990e42\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.941622 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfhnq\" (UniqueName: \"kubernetes.io/projected/2b99942b-b619-4a73-bae0-b3cd34990e42-kube-api-access-kfhnq\") pod \"cluster-logging-operator-ff9846bd-lv57g\" (UID: \"2b99942b-b619-4a73-bae0-b3cd34990e42\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" Nov 24 00:36:10 crc kubenswrapper[4888]: I1124 00:36:10.968801 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfhnq\" (UniqueName: \"kubernetes.io/projected/2b99942b-b619-4a73-bae0-b3cd34990e42-kube-api-access-kfhnq\") pod \"cluster-logging-operator-ff9846bd-lv57g\" (UID: \"2b99942b-b619-4a73-bae0-b3cd34990e42\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" Nov 24 00:36:11 crc kubenswrapper[4888]: I1124 00:36:11.051172 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" Nov 24 00:36:11 crc kubenswrapper[4888]: I1124 00:36:11.381298 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lv57g"] Nov 24 00:36:11 crc kubenswrapper[4888]: W1124 00:36:11.401739 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b99942b_b619_4a73_bae0_b3cd34990e42.slice/crio-e8d26ca495ae59109c5ac26fdadca813d097da195619c3998c49c5eaef52f6cc WatchSource:0}: Error finding container e8d26ca495ae59109c5ac26fdadca813d097da195619c3998c49c5eaef52f6cc: Status 404 returned error can't find the container with id e8d26ca495ae59109c5ac26fdadca813d097da195619c3998c49c5eaef52f6cc Nov 24 00:36:12 crc kubenswrapper[4888]: I1124 00:36:12.353106 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" event={"ID":"2b99942b-b619-4a73-bae0-b3cd34990e42","Type":"ContainerStarted","Data":"e8d26ca495ae59109c5ac26fdadca813d097da195619c3998c49c5eaef52f6cc"} Nov 24 00:36:15 crc kubenswrapper[4888]: I1124 00:36:15.378940 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" event={"ID":"d9fc74ad-5691-4d31-a9b8-9bab98d49121","Type":"ContainerStarted","Data":"f00aa1316834f8fa0e590830493a50710dcce93cdc61e6795d939f100547c7aa"} Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.459236 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" event={"ID":"2b99942b-b619-4a73-bae0-b3cd34990e42","Type":"ContainerStarted","Data":"a0f2fcb2202c404d931e7845b9f115287dfb80491a65885813f239d2b47eb1d7"} Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.462285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" event={"ID":"d9fc74ad-5691-4d31-a9b8-9bab98d49121","Type":"ContainerStarted","Data":"fd60e6bcb3d28da9e16bcfd368ac4d7a2569e0a4ad9c32a9f4cdd1558c784d8e"} Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.462612 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.468045 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.488373 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-lv57g" podStartSLOduration=2.555522859 podStartE2EDuration="13.48835037s" podCreationTimestamp="2025-11-24 00:36:10 +0000 UTC" firstStartedPulling="2025-11-24 00:36:11.405246423 +0000 UTC m=+673.987930467" lastFinishedPulling="2025-11-24 00:36:22.338073934 +0000 UTC m=+684.920757978" observedRunningTime="2025-11-24 00:36:23.485297134 +0000 UTC m=+686.067981178" watchObservedRunningTime="2025-11-24 00:36:23.48835037 +0000 UTC m=+686.071034414" Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.521639 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-648dc49c56-gnzhc" podStartSLOduration=2.757620906 podStartE2EDuration="16.521602524s" podCreationTimestamp="2025-11-24 00:36:07 +0000 UTC" firstStartedPulling="2025-11-24 00:36:08.641527644 +0000 UTC m=+671.224211688" lastFinishedPulling="2025-11-24 00:36:22.405509262 +0000 UTC m=+684.988193306" observedRunningTime="2025-11-24 00:36:23.513471006 +0000 UTC m=+686.096155070" watchObservedRunningTime="2025-11-24 00:36:23.521602524 +0000 UTC m=+686.104286608" Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.592023 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:36:23 crc kubenswrapper[4888]: I1124 00:36:23.592113 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.971156 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.972453 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.974673 4888 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-64sqq" Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.975769 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.981902 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 00:36:27 crc kubenswrapper[4888]: I1124 00:36:27.981985 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.121654 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.121745 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79png\" (UniqueName: \"kubernetes.io/projected/6cb0f0d7-48ea-4694-ae4a-59dec8bf738e-kube-api-access-79png\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.223074 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.223502 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79png\" (UniqueName: \"kubernetes.io/projected/6cb0f0d7-48ea-4694-ae4a-59dec8bf738e-kube-api-access-79png\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.227971 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.228031 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7d6b0556384b7379842ce94bf24e0bbe9b44a152b804e7c7e5b55b100dc7e645/globalmount\"" pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.252422 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-221ffacb-24cb-47eb-b8a6-fca8ee1ef447\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.252970 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79png\" (UniqueName: \"kubernetes.io/projected/6cb0f0d7-48ea-4694-ae4a-59dec8bf738e-kube-api-access-79png\") pod \"minio\" (UID: \"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e\") " pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.289494 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 00:36:28 crc kubenswrapper[4888]: I1124 00:36:28.636488 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 00:36:29 crc kubenswrapper[4888]: I1124 00:36:29.522927 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e","Type":"ContainerStarted","Data":"d01455e8522398cc65b4a412057628deb9277645de933e53d040c52ee4ff9dfa"} Nov 24 00:36:32 crc kubenswrapper[4888]: I1124 00:36:32.551469 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"6cb0f0d7-48ea-4694-ae4a-59dec8bf738e","Type":"ContainerStarted","Data":"9af0fc01e539e9b53d510a8610e371d67cbe83257de3c47306277fb57aca2d7e"} Nov 24 00:36:32 crc kubenswrapper[4888]: I1124 00:36:32.578918 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.272481228 podStartE2EDuration="7.578881165s" podCreationTimestamp="2025-11-24 00:36:25 +0000 UTC" firstStartedPulling="2025-11-24 00:36:28.651753974 +0000 UTC m=+691.234438028" lastFinishedPulling="2025-11-24 00:36:31.958153921 +0000 UTC m=+694.540837965" observedRunningTime="2025-11-24 00:36:32.574043699 +0000 UTC m=+695.156727763" watchObservedRunningTime="2025-11-24 00:36:32.578881165 +0000 UTC m=+695.161565249" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.084550 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.086090 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.106707 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-dqjxj" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.106885 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.106972 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.107272 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.115877 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.121163 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.131784 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-config\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.131884 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.131921 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.131972 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.132008 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpp6t\" (UniqueName: \"kubernetes.io/projected/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-kube-api-access-tpp6t\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.233391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.233461 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.233526 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.233581 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpp6t\" (UniqueName: \"kubernetes.io/projected/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-kube-api-access-tpp6t\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.233647 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-config\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.235212 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.235478 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-config\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.249408 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-69xbm"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.251063 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.251867 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.252124 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.258033 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.258852 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.259041 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.271536 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-69xbm"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.287137 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpp6t\" (UniqueName: \"kubernetes.io/projected/9dfccf91-b0dd-46c1-a8a0-784d83237bbe-kube-api-access-tpp6t\") pod \"logging-loki-distributor-76cc67bf56-4kw2j\" (UID: \"9dfccf91-b0dd-46c1-a8a0-784d83237bbe\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.366231 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.367097 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.369011 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.369373 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.378610 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.431415 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439269 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439323 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439425 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tddpj\" (UniqueName: \"kubernetes.io/projected/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-kube-api-access-tddpj\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439495 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd25b\" (UniqueName: \"kubernetes.io/projected/2386f5e1-9382-4d87-b641-4372f630d9d4-kube-api-access-gd25b\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439561 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-config\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439627 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439730 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439807 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439856 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-config\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.439888 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.486362 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-7m9r2"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.487857 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.491681 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-hjsdr"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.493446 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.493674 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.493784 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-984m9" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.493869 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.493933 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.494088 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.494193 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.517712 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-7m9r2"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542534 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542575 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542597 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-rbac\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542617 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tls-secret\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542642 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542666 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542690 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-config\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542710 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542730 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542751 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542785 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542804 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542845 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tls-secret\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542868 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tddpj\" (UniqueName: \"kubernetes.io/projected/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-kube-api-access-tddpj\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542888 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542904 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tenants\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542923 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542942 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfd5s\" (UniqueName: \"kubernetes.io/projected/7000141c-3ebe-45cf-bdc1-36be274b4d8c-kube-api-access-zfd5s\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542961 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd25b\" (UniqueName: \"kubernetes.io/projected/2386f5e1-9382-4d87-b641-4372f630d9d4-kube-api-access-gd25b\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.542979 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543001 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-config\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543025 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb4ln\" (UniqueName: \"kubernetes.io/projected/8c68d551-242e-47b9-8ed7-d47f6f256d3f-kube-api-access-vb4ln\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543048 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543074 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543103 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tenants\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543137 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.543157 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-rbac\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.548924 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-hjsdr"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.550637 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-config\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.550646 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-config\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.551837 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.551993 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.552797 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.561010 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.562123 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.564783 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.578542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2386f5e1-9382-4d87-b641-4372f630d9d4-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.584847 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tddpj\" (UniqueName: \"kubernetes.io/projected/03bd362f-47ff-4c5c-8ae8-b60c0c87138b-kube-api-access-tddpj\") pod \"logging-loki-querier-5895d59bb8-69xbm\" (UID: \"03bd362f-47ff-4c5c-8ae8-b60c0c87138b\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.585949 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd25b\" (UniqueName: \"kubernetes.io/projected/2386f5e1-9382-4d87-b641-4372f630d9d4-kube-api-access-gd25b\") pod \"logging-loki-query-frontend-84558f7c9f-ntcth\" (UID: \"2386f5e1-9382-4d87-b641-4372f630d9d4\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.610591 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tenants\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645772 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-rbac\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645893 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645918 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645937 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-rbac\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645971 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tls-secret\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.645996 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646047 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646075 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646104 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tls-secret\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646155 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tenants\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646176 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646212 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfd5s\" (UniqueName: \"kubernetes.io/projected/7000141c-3ebe-45cf-bdc1-36be274b4d8c-kube-api-access-zfd5s\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.646234 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.647439 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.647513 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb4ln\" (UniqueName: \"kubernetes.io/projected/8c68d551-242e-47b9-8ed7-d47f6f256d3f-kube-api-access-vb4ln\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.647542 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.648227 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.648331 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.648986 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-rbac\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.649208 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-rbac\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.649361 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.648795 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/7000141c-3ebe-45cf-bdc1-36be274b4d8c-lokistack-gateway\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.651800 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tls-secret\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.654768 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-tenants\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.654936 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.655208 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/8c68d551-242e-47b9-8ed7-d47f6f256d3f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.655578 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tenants\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.655759 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-tls-secret\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.655904 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/7000141c-3ebe-45cf-bdc1-36be274b4d8c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.663959 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb4ln\" (UniqueName: \"kubernetes.io/projected/8c68d551-242e-47b9-8ed7-d47f6f256d3f-kube-api-access-vb4ln\") pod \"logging-loki-gateway-76846dc87d-hjsdr\" (UID: \"8c68d551-242e-47b9-8ed7-d47f6f256d3f\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.667650 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfd5s\" (UniqueName: \"kubernetes.io/projected/7000141c-3ebe-45cf-bdc1-36be274b4d8c-kube-api-access-zfd5s\") pod \"logging-loki-gateway-76846dc87d-7m9r2\" (UID: \"7000141c-3ebe-45cf-bdc1-36be274b4d8c\") " pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.686923 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.730858 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j"] Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.854507 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.857325 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:37 crc kubenswrapper[4888]: I1124 00:36:37.874675 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-69xbm"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.146093 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-7m9r2"] Nov 24 00:36:38 crc kubenswrapper[4888]: W1124 00:36:38.148226 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7000141c_3ebe_45cf_bdc1_36be274b4d8c.slice/crio-c20df827a540ea3236f65d4aa6f0c04e0d8fa45e35e7242eb08a347decbad5a9 WatchSource:0}: Error finding container c20df827a540ea3236f65d4aa6f0c04e0d8fa45e35e7242eb08a347decbad5a9: Status 404 returned error can't find the container with id c20df827a540ea3236f65d4aa6f0c04e0d8fa45e35e7242eb08a347decbad5a9 Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.154382 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.192102 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76846dc87d-hjsdr"] Nov 24 00:36:38 crc kubenswrapper[4888]: W1124 00:36:38.197474 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c68d551_242e_47b9_8ed7_d47f6f256d3f.slice/crio-52681792d512ae518f65fd9a744b34a353ecbbe2514aebbfd9346342f7e49dfb WatchSource:0}: Error finding container 52681792d512ae518f65fd9a744b34a353ecbbe2514aebbfd9346342f7e49dfb: Status 404 returned error can't find the container with id 52681792d512ae518f65fd9a744b34a353ecbbe2514aebbfd9346342f7e49dfb Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.314643 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.315942 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.319786 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.319865 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.340446 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.360320 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.362260 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.372378 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.372672 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.377236 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.452875 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.453870 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.462327 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.462635 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468535 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2fwx\" (UniqueName: \"kubernetes.io/projected/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-kube-api-access-v2fwx\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468586 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468613 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468646 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-config\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468666 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468699 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468727 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468748 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-config\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468776 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgg52\" (UniqueName: \"kubernetes.io/projected/38b36998-57a3-4b56-b3dd-16f2800890d8-kube-api-access-dgg52\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468836 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.468866 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.469182 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.469366 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.469464 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.478701 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.571883 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.571983 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572060 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572458 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-532ff3b0-ef1d-4183-a927-472d74d37605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-532ff3b0-ef1d-4183-a927-472d74d37605\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572523 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572549 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpsfx\" (UniqueName: \"kubernetes.io/projected/e8d0d952-6592-40e8-a831-5e875e1f9d4b-kube-api-access-gpsfx\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572594 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2fwx\" (UniqueName: \"kubernetes.io/projected/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-kube-api-access-v2fwx\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572623 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572647 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572670 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572705 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-config\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572726 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572762 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572781 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572806 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-config\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572848 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgg52\" (UniqueName: \"kubernetes.io/projected/38b36998-57a3-4b56-b3dd-16f2800890d8-kube-api-access-dgg52\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572873 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572892 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572911 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572937 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572957 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.572978 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.574451 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.576169 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-config\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.576351 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.576775 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b36998-57a3-4b56-b3dd-16f2800890d8-config\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.578982 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579183 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579236 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f6b2f7f3e9fb3cc6133ec2ad9f34564bdf9f0a217cca33252660848ea5cc7e90/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579488 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579505 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579548 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.579564 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/84d8bf303e15d287ad33adbd9010debf8a4b06014564f390defa8a85b5fe3b76/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.580460 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/38b36998-57a3-4b56-b3dd-16f2800890d8-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.586681 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.586729 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/53c88f62bd8219968202acdf88e339f0f7f92ba34b9bfb6b8740b2efd642bde5/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.589220 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.592869 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" event={"ID":"8c68d551-242e-47b9-8ed7-d47f6f256d3f","Type":"ContainerStarted","Data":"52681792d512ae518f65fd9a744b34a353ecbbe2514aebbfd9346342f7e49dfb"} Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.594116 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" event={"ID":"2386f5e1-9382-4d87-b641-4372f630d9d4","Type":"ContainerStarted","Data":"17b7bf6786e00be00fda3f7634f171eb99438862016d32bdd0f9c6c71709638c"} Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.594798 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2fwx\" (UniqueName: \"kubernetes.io/projected/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-kube-api-access-v2fwx\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.596798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" event={"ID":"7000141c-3ebe-45cf-bdc1-36be274b4d8c","Type":"ContainerStarted","Data":"c20df827a540ea3236f65d4aa6f0c04e0d8fa45e35e7242eb08a347decbad5a9"} Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.598243 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" event={"ID":"9dfccf91-b0dd-46c1-a8a0-784d83237bbe","Type":"ContainerStarted","Data":"a236d21c0e5a4044f34b97f53c6ec34c35e689d05a1e75bf3e08d228db0a218c"} Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.600444 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/158da35a-1b60-4ef6-b46a-fd8d3e8c5c51-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.609218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" event={"ID":"03bd362f-47ff-4c5c-8ae8-b60c0c87138b","Type":"ContainerStarted","Data":"7fd5d2b61bbec105b261f1bea030ea89818244740a9f9ef2a9383eb4ac5c1fbe"} Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.618277 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgg52\" (UniqueName: \"kubernetes.io/projected/38b36998-57a3-4b56-b3dd-16f2800890d8-kube-api-access-dgg52\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.620517 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58669646-0aea-4b22-95ce-4d832ad3ec9a\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.623086 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7810fbac-be67-45d8-9b57-e6e02d6cdd93\") pod \"logging-loki-ingester-0\" (UID: \"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51\") " pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.628133 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d04975bf-169a-4a17-90f0-0e1861aceeff\") pod \"logging-loki-compactor-0\" (UID: \"38b36998-57a3-4b56-b3dd-16f2800890d8\") " pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.648547 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675423 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675482 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675525 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-532ff3b0-ef1d-4183-a927-472d74d37605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-532ff3b0-ef1d-4183-a927-472d74d37605\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675573 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675592 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpsfx\" (UniqueName: \"kubernetes.io/projected/e8d0d952-6592-40e8-a831-5e875e1f9d4b-kube-api-access-gpsfx\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.675719 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.678981 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-config\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.680331 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.683442 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.683503 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.683526 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-532ff3b0-ef1d-4183-a927-472d74d37605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-532ff3b0-ef1d-4183-a927-472d74d37605\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/673916217c5025c0519eec4b6ba2fafbe2f45cc17591dfff0a47d476e51b56f2/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.686786 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.690771 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/e8d0d952-6592-40e8-a831-5e875e1f9d4b-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.694153 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpsfx\" (UniqueName: \"kubernetes.io/projected/e8d0d952-6592-40e8-a831-5e875e1f9d4b-kube-api-access-gpsfx\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.701066 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.706261 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-532ff3b0-ef1d-4183-a927-472d74d37605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-532ff3b0-ef1d-4183-a927-472d74d37605\") pod \"logging-loki-index-gateway-0\" (UID: \"e8d0d952-6592-40e8-a831-5e875e1f9d4b\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.779376 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.904065 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: W1124 00:36:38.915589 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod158da35a_1b60_4ef6_b46a_fd8d3e8c5c51.slice/crio-0c9a2b04689c70992d19637122da6071e307403a460796b5443f8db26378f628 WatchSource:0}: Error finding container 0c9a2b04689c70992d19637122da6071e307403a460796b5443f8db26378f628: Status 404 returned error can't find the container with id 0c9a2b04689c70992d19637122da6071e307403a460796b5443f8db26378f628 Nov 24 00:36:38 crc kubenswrapper[4888]: I1124 00:36:38.947585 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 24 00:36:38 crc kubenswrapper[4888]: W1124 00:36:38.956466 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38b36998_57a3_4b56_b3dd_16f2800890d8.slice/crio-a46ef61f6da5f8dd8c19e23e969856add6ac83db829005de6d94bb26c6e5cde3 WatchSource:0}: Error finding container a46ef61f6da5f8dd8c19e23e969856add6ac83db829005de6d94bb26c6e5cde3: Status 404 returned error can't find the container with id a46ef61f6da5f8dd8c19e23e969856add6ac83db829005de6d94bb26c6e5cde3 Nov 24 00:36:39 crc kubenswrapper[4888]: I1124 00:36:39.014458 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 24 00:36:39 crc kubenswrapper[4888]: W1124 00:36:39.019427 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8d0d952_6592_40e8_a831_5e875e1f9d4b.slice/crio-ec9f9fec27ad5408576a4ceb8bfc69319d79198156104806f8a70dd345d9682c WatchSource:0}: Error finding container ec9f9fec27ad5408576a4ceb8bfc69319d79198156104806f8a70dd345d9682c: Status 404 returned error can't find the container with id ec9f9fec27ad5408576a4ceb8bfc69319d79198156104806f8a70dd345d9682c Nov 24 00:36:39 crc kubenswrapper[4888]: I1124 00:36:39.619851 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"38b36998-57a3-4b56-b3dd-16f2800890d8","Type":"ContainerStarted","Data":"a46ef61f6da5f8dd8c19e23e969856add6ac83db829005de6d94bb26c6e5cde3"} Nov 24 00:36:39 crc kubenswrapper[4888]: I1124 00:36:39.621676 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51","Type":"ContainerStarted","Data":"0c9a2b04689c70992d19637122da6071e307403a460796b5443f8db26378f628"} Nov 24 00:36:39 crc kubenswrapper[4888]: I1124 00:36:39.623251 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"e8d0d952-6592-40e8-a831-5e875e1f9d4b","Type":"ContainerStarted","Data":"ec9f9fec27ad5408576a4ceb8bfc69319d79198156104806f8a70dd345d9682c"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.659314 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" event={"ID":"8c68d551-242e-47b9-8ed7-d47f6f256d3f","Type":"ContainerStarted","Data":"4c0c04e037d91398c4186a2e2638516f743da6acc7a34a553fd5eed3874f9ca7"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.662389 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" event={"ID":"2386f5e1-9382-4d87-b641-4372f630d9d4","Type":"ContainerStarted","Data":"2b625daf974dc55e87fe2b1d57b349f9e4f3293e9a154a3f9a3215ef7d8c36a0"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.662536 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.664962 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"158da35a-1b60-4ef6-b46a-fd8d3e8c5c51","Type":"ContainerStarted","Data":"b466f5d25620b10a59e301baab5c9990bd4bb8760fd6379fe71710b3c4e3bde7"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.665102 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.668109 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" event={"ID":"7000141c-3ebe-45cf-bdc1-36be274b4d8c","Type":"ContainerStarted","Data":"e547dc066cba4e75d36f64833863a4135106081c2e08d796ed39c71a46728786"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.670547 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"e8d0d952-6592-40e8-a831-5e875e1f9d4b","Type":"ContainerStarted","Data":"9076b754e138ea7010eae123b8314f286134bea6e7e9189cce58f7fed9a0eb29"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.670696 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.672892 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" event={"ID":"9dfccf91-b0dd-46c1-a8a0-784d83237bbe","Type":"ContainerStarted","Data":"88e72536bf2ec4525c94e7c52f3352a1bfe5e2cd6ee92cbb73069ec54f68a633"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.673016 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.675303 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"38b36998-57a3-4b56-b3dd-16f2800890d8","Type":"ContainerStarted","Data":"37c39272a08f380df7f49df833b94af9ac0f79de3d961a396c3e848a957e28a6"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.678443 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" event={"ID":"03bd362f-47ff-4c5c-8ae8-b60c0c87138b","Type":"ContainerStarted","Data":"32ff4a0545c816993ab5fb06feb3f1c88fe52c04cd58f6316772747c3c2ffaa3"} Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.678655 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.695544 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" podStartSLOduration=2.036866679 podStartE2EDuration="5.695524209s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:38.157627534 +0000 UTC m=+700.740311578" lastFinishedPulling="2025-11-24 00:36:41.816285064 +0000 UTC m=+704.398969108" observedRunningTime="2025-11-24 00:36:42.689875971 +0000 UTC m=+705.272560015" watchObservedRunningTime="2025-11-24 00:36:42.695524209 +0000 UTC m=+705.278208263" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.728448 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" podStartSLOduration=1.605983127 podStartE2EDuration="5.728415743s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:37.746181248 +0000 UTC m=+700.328865292" lastFinishedPulling="2025-11-24 00:36:41.868613874 +0000 UTC m=+704.451297908" observedRunningTime="2025-11-24 00:36:42.721505149 +0000 UTC m=+705.304189193" watchObservedRunningTime="2025-11-24 00:36:42.728415743 +0000 UTC m=+705.311099797" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.805576 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.022188134 podStartE2EDuration="5.805527649s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:39.04008616 +0000 UTC m=+701.622770204" lastFinishedPulling="2025-11-24 00:36:41.823425645 +0000 UTC m=+704.406109719" observedRunningTime="2025-11-24 00:36:42.774044195 +0000 UTC m=+705.356728319" watchObservedRunningTime="2025-11-24 00:36:42.805527649 +0000 UTC m=+705.388211773" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.823534 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" podStartSLOduration=1.925636565 podStartE2EDuration="5.823503514s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:37.936702449 +0000 UTC m=+700.519386493" lastFinishedPulling="2025-11-24 00:36:41.834569358 +0000 UTC m=+704.417253442" observedRunningTime="2025-11-24 00:36:42.819300926 +0000 UTC m=+705.401984990" watchObservedRunningTime="2025-11-24 00:36:42.823503514 +0000 UTC m=+705.406187558" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.849226 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.025882998 podStartE2EDuration="5.849198876s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:38.92295163 +0000 UTC m=+701.505635674" lastFinishedPulling="2025-11-24 00:36:41.746267508 +0000 UTC m=+704.328951552" observedRunningTime="2025-11-24 00:36:42.84721106 +0000 UTC m=+705.429895104" watchObservedRunningTime="2025-11-24 00:36:42.849198876 +0000 UTC m=+705.431882920" Nov 24 00:36:42 crc kubenswrapper[4888]: I1124 00:36:42.876452 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=2.999043664 podStartE2EDuration="5.87642864s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:38.960372461 +0000 UTC m=+701.543056505" lastFinishedPulling="2025-11-24 00:36:41.837757427 +0000 UTC m=+704.420441481" observedRunningTime="2025-11-24 00:36:42.872321785 +0000 UTC m=+705.455005829" watchObservedRunningTime="2025-11-24 00:36:42.87642864 +0000 UTC m=+705.459112704" Nov 24 00:36:43 crc kubenswrapper[4888]: I1124 00:36:43.688981 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.723959 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" event={"ID":"8c68d551-242e-47b9-8ed7-d47f6f256d3f","Type":"ContainerStarted","Data":"c075f1d44d747a668fda4bf3b2e8c88bd4e3e3173a8d44f7418da610ae8f5a1d"} Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.724651 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.727744 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" event={"ID":"7000141c-3ebe-45cf-bdc1-36be274b4d8c","Type":"ContainerStarted","Data":"608b18bb1ed8ab0bdf22e4539fef57d59de1859ef52c91e0bc459a102d6083e1"} Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.728100 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.728174 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.749764 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.753120 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.758800 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.804857 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-76846dc87d-7m9r2" podStartSLOduration=2.016578168 podStartE2EDuration="10.804827141s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:38.154780734 +0000 UTC m=+700.737464768" lastFinishedPulling="2025-11-24 00:36:46.943029697 +0000 UTC m=+709.525713741" observedRunningTime="2025-11-24 00:36:47.800535621 +0000 UTC m=+710.383219685" watchObservedRunningTime="2025-11-24 00:36:47.804827141 +0000 UTC m=+710.387511185" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.805963 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" podStartSLOduration=2.073151967 podStartE2EDuration="10.805954973s" podCreationTimestamp="2025-11-24 00:36:37 +0000 UTC" firstStartedPulling="2025-11-24 00:36:38.205683104 +0000 UTC m=+700.788367148" lastFinishedPulling="2025-11-24 00:36:46.93848611 +0000 UTC m=+709.521170154" observedRunningTime="2025-11-24 00:36:47.769376276 +0000 UTC m=+710.352060360" watchObservedRunningTime="2025-11-24 00:36:47.805954973 +0000 UTC m=+710.388639017" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.857967 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:47 crc kubenswrapper[4888]: I1124 00:36:47.873248 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76846dc87d-hjsdr" Nov 24 00:36:53 crc kubenswrapper[4888]: I1124 00:36:53.591583 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:36:53 crc kubenswrapper[4888]: I1124 00:36:53.592441 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:36:57 crc kubenswrapper[4888]: I1124 00:36:57.438025 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-4kw2j" Nov 24 00:36:57 crc kubenswrapper[4888]: I1124 00:36:57.619095 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-69xbm" Nov 24 00:36:57 crc kubenswrapper[4888]: I1124 00:36:57.697330 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ntcth" Nov 24 00:36:58 crc kubenswrapper[4888]: I1124 00:36:58.659069 4888 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 24 00:36:58 crc kubenswrapper[4888]: I1124 00:36:58.659148 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="158da35a-1b60-4ef6-b46a-fd8d3e8c5c51" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 00:36:58 crc kubenswrapper[4888]: I1124 00:36:58.708654 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Nov 24 00:36:58 crc kubenswrapper[4888]: I1124 00:36:58.795948 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Nov 24 00:37:08 crc kubenswrapper[4888]: I1124 00:37:08.663487 4888 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 24 00:37:08 crc kubenswrapper[4888]: I1124 00:37:08.664535 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="158da35a-1b60-4ef6-b46a-fd8d3e8c5c51" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 00:37:18 crc kubenswrapper[4888]: I1124 00:37:18.659071 4888 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 24 00:37:18 crc kubenswrapper[4888]: I1124 00:37:18.660037 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="158da35a-1b60-4ef6-b46a-fd8d3e8c5c51" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 00:37:23 crc kubenswrapper[4888]: I1124 00:37:23.592116 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:37:23 crc kubenswrapper[4888]: I1124 00:37:23.592959 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:37:23 crc kubenswrapper[4888]: I1124 00:37:23.593057 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:37:23 crc kubenswrapper[4888]: I1124 00:37:23.594383 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:37:23 crc kubenswrapper[4888]: I1124 00:37:23.594526 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4" gracePeriod=600 Nov 24 00:37:24 crc kubenswrapper[4888]: I1124 00:37:24.061678 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4" exitCode=0 Nov 24 00:37:24 crc kubenswrapper[4888]: I1124 00:37:24.061761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4"} Nov 24 00:37:24 crc kubenswrapper[4888]: I1124 00:37:24.062147 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778"} Nov 24 00:37:24 crc kubenswrapper[4888]: I1124 00:37:24.062187 4888 scope.go:117] "RemoveContainer" containerID="c8027ea2485836e43257a67dc09abd5c4cc0a6bf956b536ceebf183b3be1fbde" Nov 24 00:37:26 crc kubenswrapper[4888]: I1124 00:37:26.474330 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:37:26 crc kubenswrapper[4888]: I1124 00:37:26.475337 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" podUID="c724c643-a37d-41a2-800c-22f0c308f331" containerName="controller-manager" containerID="cri-o://c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86" gracePeriod=30 Nov 24 00:37:26 crc kubenswrapper[4888]: I1124 00:37:26.567802 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:37:26 crc kubenswrapper[4888]: I1124 00:37:26.568183 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" podUID="0af5b51e-2144-440f-8d93-dce288248a98" containerName="route-controller-manager" containerID="cri-o://a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d" gracePeriod=30 Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.031122 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.087602 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.088788 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles\") pod \"c724c643-a37d-41a2-800c-22f0c308f331\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.088985 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca\") pod \"c724c643-a37d-41a2-800c-22f0c308f331\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.089024 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert\") pod \"c724c643-a37d-41a2-800c-22f0c308f331\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.089044 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config\") pod \"c724c643-a37d-41a2-800c-22f0c308f331\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.089112 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4c7c\" (UniqueName: \"kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c\") pod \"c724c643-a37d-41a2-800c-22f0c308f331\" (UID: \"c724c643-a37d-41a2-800c-22f0c308f331\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.091215 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca" (OuterVolumeSpecName: "client-ca") pod "c724c643-a37d-41a2-800c-22f0c308f331" (UID: "c724c643-a37d-41a2-800c-22f0c308f331"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.091867 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c724c643-a37d-41a2-800c-22f0c308f331" (UID: "c724c643-a37d-41a2-800c-22f0c308f331"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.092978 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config" (OuterVolumeSpecName: "config") pod "c724c643-a37d-41a2-800c-22f0c308f331" (UID: "c724c643-a37d-41a2-800c-22f0c308f331"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.095330 4888 generic.go:334] "Generic (PLEG): container finished" podID="c724c643-a37d-41a2-800c-22f0c308f331" containerID="c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86" exitCode=0 Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.095408 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" event={"ID":"c724c643-a37d-41a2-800c-22f0c308f331","Type":"ContainerDied","Data":"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86"} Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.095449 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" event={"ID":"c724c643-a37d-41a2-800c-22f0c308f331","Type":"ContainerDied","Data":"38020d5dc456968cbd2bc2da7ba1ed57815c415bb13257e40eeaed5084eabb9c"} Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.095477 4888 scope.go:117] "RemoveContainer" containerID="c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.095636 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ff8k7" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.096064 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c" (OuterVolumeSpecName: "kube-api-access-n4c7c") pod "c724c643-a37d-41a2-800c-22f0c308f331" (UID: "c724c643-a37d-41a2-800c-22f0c308f331"). InnerVolumeSpecName "kube-api-access-n4c7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.097118 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c724c643-a37d-41a2-800c-22f0c308f331" (UID: "c724c643-a37d-41a2-800c-22f0c308f331"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.101392 4888 generic.go:334] "Generic (PLEG): container finished" podID="0af5b51e-2144-440f-8d93-dce288248a98" containerID="a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d" exitCode=0 Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.101445 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" event={"ID":"0af5b51e-2144-440f-8d93-dce288248a98","Type":"ContainerDied","Data":"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d"} Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.101474 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" event={"ID":"0af5b51e-2144-440f-8d93-dce288248a98","Type":"ContainerDied","Data":"ca6cb5ea4c6797f963ded47a04f6eb7e6a783ddd1a4635f446032f17e43891f2"} Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.101532 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.124753 4888 scope.go:117] "RemoveContainer" containerID="c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86" Nov 24 00:37:27 crc kubenswrapper[4888]: E1124 00:37:27.128090 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86\": container with ID starting with c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86 not found: ID does not exist" containerID="c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.128143 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86"} err="failed to get container status \"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86\": rpc error: code = NotFound desc = could not find container \"c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86\": container with ID starting with c018417821bb0054cc2be02ac6c8e6eea08177c5eb13d1233231e6cc11bb5b86 not found: ID does not exist" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.128205 4888 scope.go:117] "RemoveContainer" containerID="a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.147024 4888 scope.go:117] "RemoveContainer" containerID="a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d" Nov 24 00:37:27 crc kubenswrapper[4888]: E1124 00:37:27.148420 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d\": container with ID starting with a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d not found: ID does not exist" containerID="a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.148479 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d"} err="failed to get container status \"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d\": rpc error: code = NotFound desc = could not find container \"a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d\": container with ID starting with a1e692dc7e66c207699e856ba1af4b04d0652299335e28a05c33171d5c996f9d not found: ID does not exist" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190022 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mk98\" (UniqueName: \"kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98\") pod \"0af5b51e-2144-440f-8d93-dce288248a98\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190132 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config\") pod \"0af5b51e-2144-440f-8d93-dce288248a98\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190255 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert\") pod \"0af5b51e-2144-440f-8d93-dce288248a98\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190427 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca\") pod \"0af5b51e-2144-440f-8d93-dce288248a98\" (UID: \"0af5b51e-2144-440f-8d93-dce288248a98\") " Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190794 4888 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190838 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c724c643-a37d-41a2-800c-22f0c308f331-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190847 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190856 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4c7c\" (UniqueName: \"kubernetes.io/projected/c724c643-a37d-41a2-800c-22f0c308f331-kube-api-access-n4c7c\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.190868 4888 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c724c643-a37d-41a2-800c-22f0c308f331-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.191251 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config" (OuterVolumeSpecName: "config") pod "0af5b51e-2144-440f-8d93-dce288248a98" (UID: "0af5b51e-2144-440f-8d93-dce288248a98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.191461 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca" (OuterVolumeSpecName: "client-ca") pod "0af5b51e-2144-440f-8d93-dce288248a98" (UID: "0af5b51e-2144-440f-8d93-dce288248a98"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.194844 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0af5b51e-2144-440f-8d93-dce288248a98" (UID: "0af5b51e-2144-440f-8d93-dce288248a98"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.196534 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98" (OuterVolumeSpecName: "kube-api-access-9mk98") pod "0af5b51e-2144-440f-8d93-dce288248a98" (UID: "0af5b51e-2144-440f-8d93-dce288248a98"). InnerVolumeSpecName "kube-api-access-9mk98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.274385 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5bf887fb54-mg2p4"] Nov 24 00:37:27 crc kubenswrapper[4888]: E1124 00:37:27.274708 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af5b51e-2144-440f-8d93-dce288248a98" containerName="route-controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.274721 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af5b51e-2144-440f-8d93-dce288248a98" containerName="route-controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: E1124 00:37:27.274733 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c724c643-a37d-41a2-800c-22f0c308f331" containerName="controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.274739 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c724c643-a37d-41a2-800c-22f0c308f331" containerName="controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.274891 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af5b51e-2144-440f-8d93-dce288248a98" containerName="route-controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.274905 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c724c643-a37d-41a2-800c-22f0c308f331" containerName="controller-manager" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.275473 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.292501 4888 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0af5b51e-2144-440f-8d93-dce288248a98-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.292560 4888 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.292573 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mk98\" (UniqueName: \"kubernetes.io/projected/0af5b51e-2144-440f-8d93-dce288248a98-kube-api-access-9mk98\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.292583 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af5b51e-2144-440f-8d93-dce288248a98-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.304305 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.305237 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.313893 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.330788 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bf887fb54-mg2p4"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393399 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-client-ca\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393450 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgcl8\" (UniqueName: \"kubernetes.io/projected/f59c677e-6b4a-4b85-a191-b12406efd4bb-kube-api-access-xgcl8\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393476 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-proxy-ca-bundles\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393535 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8495d\" (UniqueName: \"kubernetes.io/projected/cf128111-dbcd-4b0a-a249-e3ad98c875ef-kube-api-access-8495d\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393694 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-config\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393745 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf128111-dbcd-4b0a-a249-e3ad98c875ef-serving-cert\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.393944 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-client-ca\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.394152 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-config\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.394359 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59c677e-6b4a-4b85-a191-b12406efd4bb-serving-cert\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.428592 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.433307 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ff8k7"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.439725 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.443652 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mc99g"] Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.495211 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-client-ca\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496180 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-config\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496348 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59c677e-6b4a-4b85-a191-b12406efd4bb-serving-cert\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496446 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-client-ca\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496528 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgcl8\" (UniqueName: \"kubernetes.io/projected/f59c677e-6b4a-4b85-a191-b12406efd4bb-kube-api-access-xgcl8\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496602 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-proxy-ca-bundles\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496687 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8495d\" (UniqueName: \"kubernetes.io/projected/cf128111-dbcd-4b0a-a249-e3ad98c875ef-kube-api-access-8495d\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496776 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-config\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.496897 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf128111-dbcd-4b0a-a249-e3ad98c875ef-serving-cert\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.498195 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-client-ca\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.498349 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-config\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.498524 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-config\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.498586 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f59c677e-6b4a-4b85-a191-b12406efd4bb-proxy-ca-bundles\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.498921 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf128111-dbcd-4b0a-a249-e3ad98c875ef-client-ca\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.500832 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf128111-dbcd-4b0a-a249-e3ad98c875ef-serving-cert\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.501522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59c677e-6b4a-4b85-a191-b12406efd4bb-serving-cert\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.518551 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8495d\" (UniqueName: \"kubernetes.io/projected/cf128111-dbcd-4b0a-a249-e3ad98c875ef-kube-api-access-8495d\") pod \"route-controller-manager-6566c8c64-vdxsr\" (UID: \"cf128111-dbcd-4b0a-a249-e3ad98c875ef\") " pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.520153 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgcl8\" (UniqueName: \"kubernetes.io/projected/f59c677e-6b4a-4b85-a191-b12406efd4bb-kube-api-access-xgcl8\") pod \"controller-manager-5bf887fb54-mg2p4\" (UID: \"f59c677e-6b4a-4b85-a191-b12406efd4bb\") " pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.589773 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:27 crc kubenswrapper[4888]: I1124 00:37:27.620943 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.047522 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr"] Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.106383 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bf887fb54-mg2p4"] Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.110178 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" event={"ID":"cf128111-dbcd-4b0a-a249-e3ad98c875ef","Type":"ContainerStarted","Data":"1f9d3e5a43677bf724ea1f35219d9822be2cb128138eb0039f4ddbc6c98a1f0d"} Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.255045 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af5b51e-2144-440f-8d93-dce288248a98" path="/var/lib/kubelet/pods/0af5b51e-2144-440f-8d93-dce288248a98/volumes" Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.256350 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c724c643-a37d-41a2-800c-22f0c308f331" path="/var/lib/kubelet/pods/c724c643-a37d-41a2-800c-22f0c308f331/volumes" Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.656828 4888 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 24 00:37:28 crc kubenswrapper[4888]: I1124 00:37:28.656899 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="158da35a-1b60-4ef6-b46a-fd8d3e8c5c51" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.122223 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" event={"ID":"cf128111-dbcd-4b0a-a249-e3ad98c875ef","Type":"ContainerStarted","Data":"6beae508f72a7b9825261d025af0b137fc21050a6b670676d544935bb2c46c4a"} Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.122684 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.124357 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" event={"ID":"f59c677e-6b4a-4b85-a191-b12406efd4bb","Type":"ContainerStarted","Data":"9230646c426f3e16597584b72adbef555d39da0a0ebfd329560726f69e911dd2"} Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.124392 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" event={"ID":"f59c677e-6b4a-4b85-a191-b12406efd4bb","Type":"ContainerStarted","Data":"bc8c0adaf1b940f74cae4f1db2f39379b083c0a22a431b4fa155e4e6548bb998"} Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.124606 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.128853 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.130462 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.164556 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6566c8c64-vdxsr" podStartSLOduration=2.164523353 podStartE2EDuration="2.164523353s" podCreationTimestamp="2025-11-24 00:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:37:29.152413863 +0000 UTC m=+751.735097907" watchObservedRunningTime="2025-11-24 00:37:29.164523353 +0000 UTC m=+751.747207397" Nov 24 00:37:29 crc kubenswrapper[4888]: I1124 00:37:29.194902 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5bf887fb54-mg2p4" podStartSLOduration=2.194788163 podStartE2EDuration="2.194788163s" podCreationTimestamp="2025-11-24 00:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:37:29.191651505 +0000 UTC m=+751.774335549" watchObservedRunningTime="2025-11-24 00:37:29.194788163 +0000 UTC m=+751.777472207" Nov 24 00:37:32 crc kubenswrapper[4888]: I1124 00:37:32.862725 4888 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 00:37:38 crc kubenswrapper[4888]: I1124 00:37:38.663094 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.165416 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.168034 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.182414 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.293944 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.294121 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d864p\" (UniqueName: \"kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.294189 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.395947 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.396095 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d864p\" (UniqueName: \"kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.396165 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.396612 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.396676 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.421650 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d864p\" (UniqueName: \"kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p\") pod \"community-operators-v2vcb\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.491229 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:44 crc kubenswrapper[4888]: I1124 00:37:44.962342 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:45 crc kubenswrapper[4888]: I1124 00:37:45.264006 4888 generic.go:334] "Generic (PLEG): container finished" podID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerID="0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43" exitCode=0 Nov 24 00:37:45 crc kubenswrapper[4888]: I1124 00:37:45.265180 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerDied","Data":"0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43"} Nov 24 00:37:45 crc kubenswrapper[4888]: I1124 00:37:45.265290 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerStarted","Data":"2699e6903af066ddb02ce60950a7ed20428c2ee80fefab9d4bb0178b45e48a09"} Nov 24 00:37:46 crc kubenswrapper[4888]: I1124 00:37:46.296323 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerStarted","Data":"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3"} Nov 24 00:37:47 crc kubenswrapper[4888]: I1124 00:37:47.310769 4888 generic.go:334] "Generic (PLEG): container finished" podID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerID="bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3" exitCode=0 Nov 24 00:37:47 crc kubenswrapper[4888]: I1124 00:37:47.310997 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerDied","Data":"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3"} Nov 24 00:37:48 crc kubenswrapper[4888]: I1124 00:37:48.324484 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerStarted","Data":"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876"} Nov 24 00:37:48 crc kubenswrapper[4888]: I1124 00:37:48.353141 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v2vcb" podStartSLOduration=1.823012316 podStartE2EDuration="4.35312028s" podCreationTimestamp="2025-11-24 00:37:44 +0000 UTC" firstStartedPulling="2025-11-24 00:37:45.266687262 +0000 UTC m=+767.849371296" lastFinishedPulling="2025-11-24 00:37:47.796795176 +0000 UTC m=+770.379479260" observedRunningTime="2025-11-24 00:37:48.349021825 +0000 UTC m=+770.931705899" watchObservedRunningTime="2025-11-24 00:37:48.35312028 +0000 UTC m=+770.935804324" Nov 24 00:37:54 crc kubenswrapper[4888]: I1124 00:37:54.493111 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:54 crc kubenswrapper[4888]: I1124 00:37:54.493859 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:54 crc kubenswrapper[4888]: I1124 00:37:54.568362 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:55 crc kubenswrapper[4888]: I1124 00:37:55.467351 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:55 crc kubenswrapper[4888]: I1124 00:37:55.537199 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.191699 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-pwp9w"] Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.193024 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.198579 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-xfrbs" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.199158 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.215490 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-pwp9w"] Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.215893 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.216847 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.216969 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.223903 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.295009 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-pwp9w"] Nov 24 00:37:57 crc kubenswrapper[4888]: E1124 00:37:57.295592 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-d6sdh metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-pwp9w" podUID="7312cd78-9bf2-4089-861f-37062cf2c728" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.350960 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351069 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351203 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351364 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351473 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351532 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.351553 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.352041 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.352109 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6sdh\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.352140 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.352165 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.361152 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.363632 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.376054 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.409213 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.409960 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v2vcb" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="registry-server" containerID="cri-o://48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876" gracePeriod=2 Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.427587 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454245 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454302 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454339 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454425 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454468 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzg7z\" (UniqueName: \"kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454506 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454534 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454561 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454583 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454658 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6sdh\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454713 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454740 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.454866 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.455335 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.455929 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.455955 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: E1124 00:37:57.456067 4888 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Nov 24 00:37:57 crc kubenswrapper[4888]: E1124 00:37:57.456156 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver podName:7312cd78-9bf2-4089-861f-37062cf2c728 nodeName:}" failed. No retries permitted until 2025-11-24 00:37:57.956130804 +0000 UTC m=+780.538814868 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver") pod "collector-pwp9w" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728") : secret "collector-syslog-receiver" not found Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.458615 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.467596 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.470522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.476542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.483627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6sdh\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.484317 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557307 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557379 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6sdh\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557413 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557484 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557578 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557634 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557703 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557734 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557780 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.557827 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.558030 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.558107 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzg7z\" (UniqueName: \"kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.558157 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.558798 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.559186 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.559240 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir" (OuterVolumeSpecName: "datadir") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.559238 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.559308 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config" (OuterVolumeSpecName: "config") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.559567 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.560001 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.564105 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token" (OuterVolumeSpecName: "sa-token") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.567072 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp" (OuterVolumeSpecName: "tmp") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.567360 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh" (OuterVolumeSpecName: "kube-api-access-d6sdh") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "kube-api-access-d6sdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.567882 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics" (OuterVolumeSpecName: "metrics") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.580389 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzg7z\" (UniqueName: \"kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z\") pod \"redhat-marketplace-gpvsf\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.587781 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token" (OuterVolumeSpecName: "collector-token") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.660539 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661087 4888 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661100 4888 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-entrypoint\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661112 4888 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7312cd78-9bf2-4089-861f-37062cf2c728-datadir\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661121 4888 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7312cd78-9bf2-4089-861f-37062cf2c728-tmp\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661129 4888 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661138 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6sdh\" (UniqueName: \"kubernetes.io/projected/7312cd78-9bf2-4089-861f-37062cf2c728-kube-api-access-d6sdh\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661153 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661163 4888 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7312cd78-9bf2-4089-861f-37062cf2c728-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.661237 4888 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-token\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.685315 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.891297 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.967862 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content\") pod \"34f172d7-c628-459d-9d8f-b664a18c08c5\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.968005 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d864p\" (UniqueName: \"kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p\") pod \"34f172d7-c628-459d-9d8f-b664a18c08c5\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.968210 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities\") pod \"34f172d7-c628-459d-9d8f-b664a18c08c5\" (UID: \"34f172d7-c628-459d-9d8f-b664a18c08c5\") " Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.968592 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.969063 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities" (OuterVolumeSpecName: "utilities") pod "34f172d7-c628-459d-9d8f-b664a18c08c5" (UID: "34f172d7-c628-459d-9d8f-b664a18c08c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.973836 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") pod \"collector-pwp9w\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " pod="openshift-logging/collector-pwp9w" Nov 24 00:37:57 crc kubenswrapper[4888]: I1124 00:37:57.974300 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p" (OuterVolumeSpecName: "kube-api-access-d864p") pod "34f172d7-c628-459d-9d8f-b664a18c08c5" (UID: "34f172d7-c628-459d-9d8f-b664a18c08c5"). InnerVolumeSpecName "kube-api-access-d864p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.026315 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34f172d7-c628-459d-9d8f-b664a18c08c5" (UID: "34f172d7-c628-459d-9d8f-b664a18c08c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.070222 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") pod \"7312cd78-9bf2-4089-861f-37062cf2c728\" (UID: \"7312cd78-9bf2-4089-861f-37062cf2c728\") " Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.070731 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d864p\" (UniqueName: \"kubernetes.io/projected/34f172d7-c628-459d-9d8f-b664a18c08c5-kube-api-access-d864p\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.070755 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.070771 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f172d7-c628-459d-9d8f-b664a18c08c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.073996 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "7312cd78-9bf2-4089-861f-37062cf2c728" (UID: "7312cd78-9bf2-4089-861f-37062cf2c728"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.162689 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:37:58 crc kubenswrapper[4888]: W1124 00:37:58.169325 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93053d48_a14c_407e_aedd_1d7d0cd3e877.slice/crio-b1c1370e4fc65c80dc1a5dc6d3ec5af710f11c554815c5cf38fb062b4d1844be WatchSource:0}: Error finding container b1c1370e4fc65c80dc1a5dc6d3ec5af710f11c554815c5cf38fb062b4d1844be: Status 404 returned error can't find the container with id b1c1370e4fc65c80dc1a5dc6d3ec5af710f11c554815c5cf38fb062b4d1844be Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.171667 4888 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7312cd78-9bf2-4089-861f-37062cf2c728-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.416480 4888 generic.go:334] "Generic (PLEG): container finished" podID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerID="48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876" exitCode=0 Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.417005 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerDied","Data":"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876"} Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.417040 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2vcb" event={"ID":"34f172d7-c628-459d-9d8f-b664a18c08c5","Type":"ContainerDied","Data":"2699e6903af066ddb02ce60950a7ed20428c2ee80fefab9d4bb0178b45e48a09"} Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.417061 4888 scope.go:117] "RemoveContainer" containerID="48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.417216 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2vcb" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.420078 4888 generic.go:334] "Generic (PLEG): container finished" podID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerID="51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd" exitCode=0 Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.420191 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-pwp9w" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.420970 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerDied","Data":"51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd"} Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.421088 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerStarted","Data":"b1c1370e4fc65c80dc1a5dc6d3ec5af710f11c554815c5cf38fb062b4d1844be"} Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.440086 4888 scope.go:117] "RemoveContainer" containerID="bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.464015 4888 scope.go:117] "RemoveContainer" containerID="0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.495139 4888 scope.go:117] "RemoveContainer" containerID="48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876" Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.495866 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876\": container with ID starting with 48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876 not found: ID does not exist" containerID="48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.495957 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876"} err="failed to get container status \"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876\": rpc error: code = NotFound desc = could not find container \"48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876\": container with ID starting with 48b48b7d171faa632d53b254cc6cc910955f57ba3984f23ddea84b1d58d81876 not found: ID does not exist" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.496031 4888 scope.go:117] "RemoveContainer" containerID="bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3" Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.496929 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3\": container with ID starting with bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3 not found: ID does not exist" containerID="bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.496971 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3"} err="failed to get container status \"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3\": rpc error: code = NotFound desc = could not find container \"bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3\": container with ID starting with bac3d4c1b76f44068834b0b9386ecda9f6232292bf6b8b68401756b89ebb0dc3 not found: ID does not exist" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.497002 4888 scope.go:117] "RemoveContainer" containerID="0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43" Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.498883 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43\": container with ID starting with 0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43 not found: ID does not exist" containerID="0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.498927 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43"} err="failed to get container status \"0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43\": rpc error: code = NotFound desc = could not find container \"0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43\": container with ID starting with 0cff5adf41cb2bdb2dc4d2c37665a3456ed7ce9d19301d0bb550b6613c659a43 not found: ID does not exist" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.501610 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-pwp9w"] Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.512969 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-pwp9w"] Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.520486 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-gmbz6"] Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.521042 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="extract-content" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.521084 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="extract-content" Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.521104 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="extract-utilities" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.521112 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="extract-utilities" Nov 24 00:37:58 crc kubenswrapper[4888]: E1124 00:37:58.521132 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="registry-server" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.521140 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="registry-server" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.521261 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" containerName="registry-server" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.522026 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.528382 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.528743 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.528923 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-xfrbs" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.529061 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.533955 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.535094 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.549111 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.557159 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-gmbz6"] Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.565851 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v2vcb"] Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679527 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679591 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-syslog-receiver\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679650 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/61cf37cb-90a9-4b6a-8252-30a564f91b0d-tmp\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679732 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/61cf37cb-90a9-4b6a-8252-30a564f91b0d-datadir\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679774 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.679799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-entrypoint\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.680027 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-trusted-ca\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.680107 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-sa-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.680300 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-metrics\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.680424 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config-openshift-service-cacrt\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.680469 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62t7c\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-kube-api-access-62t7c\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782716 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-metrics\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782786 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config-openshift-service-cacrt\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782806 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62t7c\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-kube-api-access-62t7c\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782858 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782878 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-syslog-receiver\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782896 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/61cf37cb-90a9-4b6a-8252-30a564f91b0d-tmp\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.782930 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/61cf37cb-90a9-4b6a-8252-30a564f91b0d-datadir\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.783146 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.783167 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-entrypoint\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.783201 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-trusted-ca\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.783189 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/61cf37cb-90a9-4b6a-8252-30a564f91b0d-datadir\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.783226 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-sa-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.784206 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-trusted-ca\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.784322 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config-openshift-service-cacrt\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.785638 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-entrypoint\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.785713 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61cf37cb-90a9-4b6a-8252-30a564f91b0d-config\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.788113 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/61cf37cb-90a9-4b6a-8252-30a564f91b0d-tmp\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.789137 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-syslog-receiver\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.789668 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-collector-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.798662 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/61cf37cb-90a9-4b6a-8252-30a564f91b0d-metrics\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.801181 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-sa-token\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.811189 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62t7c\" (UniqueName: \"kubernetes.io/projected/61cf37cb-90a9-4b6a-8252-30a564f91b0d-kube-api-access-62t7c\") pod \"collector-gmbz6\" (UID: \"61cf37cb-90a9-4b6a-8252-30a564f91b0d\") " pod="openshift-logging/collector-gmbz6" Nov 24 00:37:58 crc kubenswrapper[4888]: I1124 00:37:58.878072 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gmbz6" Nov 24 00:37:59 crc kubenswrapper[4888]: I1124 00:37:59.394842 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-gmbz6"] Nov 24 00:37:59 crc kubenswrapper[4888]: W1124 00:37:59.416198 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61cf37cb_90a9_4b6a_8252_30a564f91b0d.slice/crio-6f881091d37c4bf55c5109985610825dd52793f7f516afa79b2edbae34def7ac WatchSource:0}: Error finding container 6f881091d37c4bf55c5109985610825dd52793f7f516afa79b2edbae34def7ac: Status 404 returned error can't find the container with id 6f881091d37c4bf55c5109985610825dd52793f7f516afa79b2edbae34def7ac Nov 24 00:37:59 crc kubenswrapper[4888]: I1124 00:37:59.437706 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-gmbz6" event={"ID":"61cf37cb-90a9-4b6a-8252-30a564f91b0d","Type":"ContainerStarted","Data":"6f881091d37c4bf55c5109985610825dd52793f7f516afa79b2edbae34def7ac"} Nov 24 00:37:59 crc kubenswrapper[4888]: I1124 00:37:59.441067 4888 generic.go:334] "Generic (PLEG): container finished" podID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerID="d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4" exitCode=0 Nov 24 00:37:59 crc kubenswrapper[4888]: I1124 00:37:59.441145 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerDied","Data":"d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4"} Nov 24 00:37:59 crc kubenswrapper[4888]: E1124 00:37:59.801369 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:00 crc kubenswrapper[4888]: I1124 00:38:00.260739 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f172d7-c628-459d-9d8f-b664a18c08c5" path="/var/lib/kubelet/pods/34f172d7-c628-459d-9d8f-b664a18c08c5/volumes" Nov 24 00:38:00 crc kubenswrapper[4888]: I1124 00:38:00.262475 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7312cd78-9bf2-4089-861f-37062cf2c728" path="/var/lib/kubelet/pods/7312cd78-9bf2-4089-861f-37062cf2c728/volumes" Nov 24 00:38:00 crc kubenswrapper[4888]: I1124 00:38:00.453763 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerStarted","Data":"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de"} Nov 24 00:38:00 crc kubenswrapper[4888]: I1124 00:38:00.485883 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gpvsf" podStartSLOduration=2.033941729 podStartE2EDuration="3.485859527s" podCreationTimestamp="2025-11-24 00:37:57 +0000 UTC" firstStartedPulling="2025-11-24 00:37:58.422025413 +0000 UTC m=+781.004709457" lastFinishedPulling="2025-11-24 00:37:59.873943191 +0000 UTC m=+782.456627255" observedRunningTime="2025-11-24 00:38:00.483338437 +0000 UTC m=+783.066022521" watchObservedRunningTime="2025-11-24 00:38:00.485859527 +0000 UTC m=+783.068543581" Nov 24 00:38:04 crc kubenswrapper[4888]: I1124 00:38:04.985031 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:04 crc kubenswrapper[4888]: I1124 00:38:04.991062 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.031190 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.116869 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szd7j\" (UniqueName: \"kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.117130 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.117328 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.219407 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szd7j\" (UniqueName: \"kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.219484 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.219516 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.220389 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.220388 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.247314 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szd7j\" (UniqueName: \"kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j\") pod \"certified-operators-7nn9l\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:05 crc kubenswrapper[4888]: I1124 00:38:05.358878 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:06 crc kubenswrapper[4888]: I1124 00:38:06.453611 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:06 crc kubenswrapper[4888]: I1124 00:38:06.504626 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-gmbz6" event={"ID":"61cf37cb-90a9-4b6a-8252-30a564f91b0d","Type":"ContainerStarted","Data":"0377a2b9e73d6867b6cab1dad017466f0322ad254ddee649ae0f25774fdb3941"} Nov 24 00:38:06 crc kubenswrapper[4888]: I1124 00:38:06.506504 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerStarted","Data":"05aba020a2dffd33b2ccb3de93e32fc2a6d04af30e071b57020eb7a210e147a1"} Nov 24 00:38:06 crc kubenswrapper[4888]: I1124 00:38:06.529953 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-gmbz6" podStartSLOduration=1.94509162 podStartE2EDuration="8.529884175s" podCreationTimestamp="2025-11-24 00:37:58 +0000 UTC" firstStartedPulling="2025-11-24 00:37:59.422268905 +0000 UTC m=+782.004952949" lastFinishedPulling="2025-11-24 00:38:06.00706146 +0000 UTC m=+788.589745504" observedRunningTime="2025-11-24 00:38:06.526664175 +0000 UTC m=+789.109348229" watchObservedRunningTime="2025-11-24 00:38:06.529884175 +0000 UTC m=+789.112568219" Nov 24 00:38:07 crc kubenswrapper[4888]: I1124 00:38:07.519655 4888 generic.go:334] "Generic (PLEG): container finished" podID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerID="98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a" exitCode=0 Nov 24 00:38:07 crc kubenswrapper[4888]: I1124 00:38:07.519749 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerDied","Data":"98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a"} Nov 24 00:38:07 crc kubenswrapper[4888]: I1124 00:38:07.685871 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:07 crc kubenswrapper[4888]: I1124 00:38:07.685998 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:07 crc kubenswrapper[4888]: I1124 00:38:07.756394 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:08 crc kubenswrapper[4888]: I1124 00:38:08.531951 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerStarted","Data":"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7"} Nov 24 00:38:08 crc kubenswrapper[4888]: I1124 00:38:08.604292 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:09 crc kubenswrapper[4888]: I1124 00:38:09.542080 4888 generic.go:334] "Generic (PLEG): container finished" podID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerID="1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7" exitCode=0 Nov 24 00:38:09 crc kubenswrapper[4888]: I1124 00:38:09.542144 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerDied","Data":"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7"} Nov 24 00:38:09 crc kubenswrapper[4888]: E1124 00:38:09.990450 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:10 crc kubenswrapper[4888]: I1124 00:38:10.018084 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:38:10 crc kubenswrapper[4888]: I1124 00:38:10.554544 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerStarted","Data":"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88"} Nov 24 00:38:10 crc kubenswrapper[4888]: I1124 00:38:10.554773 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gpvsf" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="registry-server" containerID="cri-o://7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de" gracePeriod=2 Nov 24 00:38:10 crc kubenswrapper[4888]: I1124 00:38:10.590801 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7nn9l" podStartSLOduration=4.141669015 podStartE2EDuration="6.590773712s" podCreationTimestamp="2025-11-24 00:38:04 +0000 UTC" firstStartedPulling="2025-11-24 00:38:07.522386932 +0000 UTC m=+790.105071006" lastFinishedPulling="2025-11-24 00:38:09.971491619 +0000 UTC m=+792.554175703" observedRunningTime="2025-11-24 00:38:10.586888803 +0000 UTC m=+793.169572857" watchObservedRunningTime="2025-11-24 00:38:10.590773712 +0000 UTC m=+793.173457756" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.069298 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.240242 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities\") pod \"93053d48-a14c-407e-aedd-1d7d0cd3e877\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.240452 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzg7z\" (UniqueName: \"kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z\") pod \"93053d48-a14c-407e-aedd-1d7d0cd3e877\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.240536 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content\") pod \"93053d48-a14c-407e-aedd-1d7d0cd3e877\" (UID: \"93053d48-a14c-407e-aedd-1d7d0cd3e877\") " Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.241732 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities" (OuterVolumeSpecName: "utilities") pod "93053d48-a14c-407e-aedd-1d7d0cd3e877" (UID: "93053d48-a14c-407e-aedd-1d7d0cd3e877"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.247108 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z" (OuterVolumeSpecName: "kube-api-access-mzg7z") pod "93053d48-a14c-407e-aedd-1d7d0cd3e877" (UID: "93053d48-a14c-407e-aedd-1d7d0cd3e877"). InnerVolumeSpecName "kube-api-access-mzg7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.281222 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93053d48-a14c-407e-aedd-1d7d0cd3e877" (UID: "93053d48-a14c-407e-aedd-1d7d0cd3e877"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.344619 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.344658 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93053d48-a14c-407e-aedd-1d7d0cd3e877-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.344672 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzg7z\" (UniqueName: \"kubernetes.io/projected/93053d48-a14c-407e-aedd-1d7d0cd3e877-kube-api-access-mzg7z\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.567347 4888 generic.go:334] "Generic (PLEG): container finished" podID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerID="7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de" exitCode=0 Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.567448 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpvsf" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.567469 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerDied","Data":"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de"} Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.567585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpvsf" event={"ID":"93053d48-a14c-407e-aedd-1d7d0cd3e877","Type":"ContainerDied","Data":"b1c1370e4fc65c80dc1a5dc6d3ec5af710f11c554815c5cf38fb062b4d1844be"} Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.567622 4888 scope.go:117] "RemoveContainer" containerID="7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.605531 4888 scope.go:117] "RemoveContainer" containerID="d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.626565 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.636327 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpvsf"] Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.653593 4888 scope.go:117] "RemoveContainer" containerID="51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.683076 4888 scope.go:117] "RemoveContainer" containerID="7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de" Nov 24 00:38:11 crc kubenswrapper[4888]: E1124 00:38:11.684028 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de\": container with ID starting with 7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de not found: ID does not exist" containerID="7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.684070 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de"} err="failed to get container status \"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de\": rpc error: code = NotFound desc = could not find container \"7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de\": container with ID starting with 7d7c14f1ab846f3d39be1baa575b3332735ed7d2259f919604a6a9b77071d9de not found: ID does not exist" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.684100 4888 scope.go:117] "RemoveContainer" containerID="d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4" Nov 24 00:38:11 crc kubenswrapper[4888]: E1124 00:38:11.684733 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4\": container with ID starting with d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4 not found: ID does not exist" containerID="d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.684953 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4"} err="failed to get container status \"d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4\": rpc error: code = NotFound desc = could not find container \"d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4\": container with ID starting with d608a0bcee17e59d6122dc8c66378ea986f56813aebca9cfeb61ebb4bef549d4 not found: ID does not exist" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.685120 4888 scope.go:117] "RemoveContainer" containerID="51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd" Nov 24 00:38:11 crc kubenswrapper[4888]: E1124 00:38:11.685651 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd\": container with ID starting with 51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd not found: ID does not exist" containerID="51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd" Nov 24 00:38:11 crc kubenswrapper[4888]: I1124 00:38:11.685679 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd"} err="failed to get container status \"51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd\": rpc error: code = NotFound desc = could not find container \"51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd\": container with ID starting with 51d9478f298c4b7bb5cb2236cbf95cef5444a3e45fde03b0c58afc237770a0cd not found: ID does not exist" Nov 24 00:38:12 crc kubenswrapper[4888]: I1124 00:38:12.263041 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" path="/var/lib/kubelet/pods/93053d48-a14c-407e-aedd-1d7d0cd3e877/volumes" Nov 24 00:38:15 crc kubenswrapper[4888]: I1124 00:38:15.359741 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:15 crc kubenswrapper[4888]: I1124 00:38:15.359854 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:15 crc kubenswrapper[4888]: I1124 00:38:15.418517 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:15 crc kubenswrapper[4888]: I1124 00:38:15.665310 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:16 crc kubenswrapper[4888]: I1124 00:38:16.818344 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:17 crc kubenswrapper[4888]: I1124 00:38:17.622321 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7nn9l" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="registry-server" containerID="cri-o://877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88" gracePeriod=2 Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.151473 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.280755 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szd7j\" (UniqueName: \"kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j\") pod \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.281447 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content\") pod \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.281735 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities\") pod \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\" (UID: \"501bf1fd-7310-4fdf-a50c-6c0ad4bee784\") " Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.283152 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities" (OuterVolumeSpecName: "utilities") pod "501bf1fd-7310-4fdf-a50c-6c0ad4bee784" (UID: "501bf1fd-7310-4fdf-a50c-6c0ad4bee784"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.292466 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j" (OuterVolumeSpecName: "kube-api-access-szd7j") pod "501bf1fd-7310-4fdf-a50c-6c0ad4bee784" (UID: "501bf1fd-7310-4fdf-a50c-6c0ad4bee784"). InnerVolumeSpecName "kube-api-access-szd7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.381811 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "501bf1fd-7310-4fdf-a50c-6c0ad4bee784" (UID: "501bf1fd-7310-4fdf-a50c-6c0ad4bee784"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.384542 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.384586 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.384605 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szd7j\" (UniqueName: \"kubernetes.io/projected/501bf1fd-7310-4fdf-a50c-6c0ad4bee784-kube-api-access-szd7j\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.643651 4888 generic.go:334] "Generic (PLEG): container finished" podID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerID="877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88" exitCode=0 Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.643731 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerDied","Data":"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88"} Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.643775 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nn9l" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.643860 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nn9l" event={"ID":"501bf1fd-7310-4fdf-a50c-6c0ad4bee784","Type":"ContainerDied","Data":"05aba020a2dffd33b2ccb3de93e32fc2a6d04af30e071b57020eb7a210e147a1"} Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.643913 4888 scope.go:117] "RemoveContainer" containerID="877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.686342 4888 scope.go:117] "RemoveContainer" containerID="1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.693365 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.706780 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7nn9l"] Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.720466 4888 scope.go:117] "RemoveContainer" containerID="98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.749114 4888 scope.go:117] "RemoveContainer" containerID="877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88" Nov 24 00:38:18 crc kubenswrapper[4888]: E1124 00:38:18.749753 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88\": container with ID starting with 877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88 not found: ID does not exist" containerID="877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.749865 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88"} err="failed to get container status \"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88\": rpc error: code = NotFound desc = could not find container \"877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88\": container with ID starting with 877b4df8caa3e37246b504b9aa07285617fe177c00bc7b4da786c47df8c4ec88 not found: ID does not exist" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.749912 4888 scope.go:117] "RemoveContainer" containerID="1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7" Nov 24 00:38:18 crc kubenswrapper[4888]: E1124 00:38:18.750425 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7\": container with ID starting with 1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7 not found: ID does not exist" containerID="1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.750476 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7"} err="failed to get container status \"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7\": rpc error: code = NotFound desc = could not find container \"1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7\": container with ID starting with 1fb08ac4dfbdab27de5a8f734399238b96af9dcca9e3389272336a1e577014f7 not found: ID does not exist" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.750513 4888 scope.go:117] "RemoveContainer" containerID="98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a" Nov 24 00:38:18 crc kubenswrapper[4888]: E1124 00:38:18.751225 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a\": container with ID starting with 98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a not found: ID does not exist" containerID="98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a" Nov 24 00:38:18 crc kubenswrapper[4888]: I1124 00:38:18.751261 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a"} err="failed to get container status \"98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a\": rpc error: code = NotFound desc = could not find container \"98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a\": container with ID starting with 98a797cc02204e8c38c40409def27550c364537e163f7783ed7431d08fac9d3a not found: ID does not exist" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.843218 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.843875 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.843908 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.843937 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="extract-content" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.843950 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="extract-content" Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.843969 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="extract-utilities" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.843982 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="extract-utilities" Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.844020 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="extract-utilities" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.844033 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="extract-utilities" Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.844055 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.844069 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: E1124 00:38:19.844089 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="extract-content" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.844101 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="extract-content" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.844377 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.844404 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="93053d48-a14c-407e-aedd-1d7d0cd3e877" containerName="registry-server" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.849493 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:19 crc kubenswrapper[4888]: I1124 00:38:19.853394 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.014551 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmp8\" (UniqueName: \"kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.014843 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.014923 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.116050 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.116099 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.116184 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmp8\" (UniqueName: \"kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.117194 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.117328 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.152356 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmp8\" (UniqueName: \"kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8\") pod \"redhat-operators-pxzqq\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: E1124 00:38:20.180864 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.184200 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.255612 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501bf1fd-7310-4fdf-a50c-6c0ad4bee784" path="/var/lib/kubelet/pods/501bf1fd-7310-4fdf-a50c-6c0ad4bee784/volumes" Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.622364 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:20 crc kubenswrapper[4888]: I1124 00:38:20.663622 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerStarted","Data":"703b687ed031a94ce23687d445a0c76b9095bed2d3d05c83b2c926829e230fc6"} Nov 24 00:38:21 crc kubenswrapper[4888]: I1124 00:38:21.673931 4888 generic.go:334] "Generic (PLEG): container finished" podID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerID="b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec" exitCode=0 Nov 24 00:38:21 crc kubenswrapper[4888]: I1124 00:38:21.674033 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerDied","Data":"b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec"} Nov 24 00:38:22 crc kubenswrapper[4888]: I1124 00:38:22.685751 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerStarted","Data":"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be"} Nov 24 00:38:23 crc kubenswrapper[4888]: I1124 00:38:23.697131 4888 generic.go:334] "Generic (PLEG): container finished" podID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerID="b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be" exitCode=0 Nov 24 00:38:23 crc kubenswrapper[4888]: I1124 00:38:23.697300 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerDied","Data":"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be"} Nov 24 00:38:24 crc kubenswrapper[4888]: I1124 00:38:24.710090 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerStarted","Data":"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46"} Nov 24 00:38:24 crc kubenswrapper[4888]: I1124 00:38:24.741864 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxzqq" podStartSLOduration=3.321842631 podStartE2EDuration="5.741843281s" podCreationTimestamp="2025-11-24 00:38:19 +0000 UTC" firstStartedPulling="2025-11-24 00:38:21.676430783 +0000 UTC m=+804.259114837" lastFinishedPulling="2025-11-24 00:38:24.096431443 +0000 UTC m=+806.679115487" observedRunningTime="2025-11-24 00:38:24.735430951 +0000 UTC m=+807.318115035" watchObservedRunningTime="2025-11-24 00:38:24.741843281 +0000 UTC m=+807.324527325" Nov 24 00:38:30 crc kubenswrapper[4888]: I1124 00:38:30.185040 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:30 crc kubenswrapper[4888]: I1124 00:38:30.185705 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:30 crc kubenswrapper[4888]: E1124 00:38:30.419695 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.276379 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pxzqq" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="registry-server" probeResult="failure" output=< Nov 24 00:38:31 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:38:31 crc kubenswrapper[4888]: > Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.484135 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x"] Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.485555 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.492316 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.505077 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x"] Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.634166 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vccv\" (UniqueName: \"kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.634249 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.634299 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.735983 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.736167 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vccv\" (UniqueName: \"kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.736233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.737565 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.737652 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.770065 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vccv\" (UniqueName: \"kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:31 crc kubenswrapper[4888]: I1124 00:38:31.805122 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:32 crc kubenswrapper[4888]: I1124 00:38:32.254383 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x"] Nov 24 00:38:32 crc kubenswrapper[4888]: I1124 00:38:32.776329 4888 generic.go:334] "Generic (PLEG): container finished" podID="e749cc09-5587-4044-878a-c3adbd64909a" containerID="49c011b470ca6681aa1c40a43727cd249970d4b33de27f80e4466e8f6aab4ffa" exitCode=0 Nov 24 00:38:32 crc kubenswrapper[4888]: I1124 00:38:32.776442 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" event={"ID":"e749cc09-5587-4044-878a-c3adbd64909a","Type":"ContainerDied","Data":"49c011b470ca6681aa1c40a43727cd249970d4b33de27f80e4466e8f6aab4ffa"} Nov 24 00:38:32 crc kubenswrapper[4888]: I1124 00:38:32.776631 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" event={"ID":"e749cc09-5587-4044-878a-c3adbd64909a","Type":"ContainerStarted","Data":"cbb5cc67ca5c08ebdacf850a7b93f50bf059545979a8010fdd46d0b1780e68ae"} Nov 24 00:38:35 crc kubenswrapper[4888]: I1124 00:38:35.802417 4888 generic.go:334] "Generic (PLEG): container finished" podID="e749cc09-5587-4044-878a-c3adbd64909a" containerID="f159511769b25fcda6bf7fbd2d464422969c8f9fd4fa0dd3528fde2fe6dd8ed0" exitCode=0 Nov 24 00:38:35 crc kubenswrapper[4888]: I1124 00:38:35.802544 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" event={"ID":"e749cc09-5587-4044-878a-c3adbd64909a","Type":"ContainerDied","Data":"f159511769b25fcda6bf7fbd2d464422969c8f9fd4fa0dd3528fde2fe6dd8ed0"} Nov 24 00:38:36 crc kubenswrapper[4888]: I1124 00:38:36.820165 4888 generic.go:334] "Generic (PLEG): container finished" podID="e749cc09-5587-4044-878a-c3adbd64909a" containerID="e92763c0c5341e8adc812cddf7c3329b4b7025bbcfaa2cc679dea5a18b4f3427" exitCode=0 Nov 24 00:38:36 crc kubenswrapper[4888]: I1124 00:38:36.820201 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" event={"ID":"e749cc09-5587-4044-878a-c3adbd64909a","Type":"ContainerDied","Data":"e92763c0c5341e8adc812cddf7c3329b4b7025bbcfaa2cc679dea5a18b4f3427"} Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.189172 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.357949 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle\") pod \"e749cc09-5587-4044-878a-c3adbd64909a\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.358023 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vccv\" (UniqueName: \"kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv\") pod \"e749cc09-5587-4044-878a-c3adbd64909a\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.358068 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util\") pod \"e749cc09-5587-4044-878a-c3adbd64909a\" (UID: \"e749cc09-5587-4044-878a-c3adbd64909a\") " Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.360370 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle" (OuterVolumeSpecName: "bundle") pod "e749cc09-5587-4044-878a-c3adbd64909a" (UID: "e749cc09-5587-4044-878a-c3adbd64909a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.365003 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv" (OuterVolumeSpecName: "kube-api-access-4vccv") pod "e749cc09-5587-4044-878a-c3adbd64909a" (UID: "e749cc09-5587-4044-878a-c3adbd64909a"). InnerVolumeSpecName "kube-api-access-4vccv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.383021 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util" (OuterVolumeSpecName: "util") pod "e749cc09-5587-4044-878a-c3adbd64909a" (UID: "e749cc09-5587-4044-878a-c3adbd64909a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.462016 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.462068 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vccv\" (UniqueName: \"kubernetes.io/projected/e749cc09-5587-4044-878a-c3adbd64909a-kube-api-access-4vccv\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.462081 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e749cc09-5587-4044-878a-c3adbd64909a-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.839937 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" event={"ID":"e749cc09-5587-4044-878a-c3adbd64909a","Type":"ContainerDied","Data":"cbb5cc67ca5c08ebdacf850a7b93f50bf059545979a8010fdd46d0b1780e68ae"} Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.840028 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbb5cc67ca5c08ebdacf850a7b93f50bf059545979a8010fdd46d0b1780e68ae" Nov 24 00:38:38 crc kubenswrapper[4888]: I1124 00:38:38.840158 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x" Nov 24 00:38:40 crc kubenswrapper[4888]: I1124 00:38:40.282935 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:40 crc kubenswrapper[4888]: I1124 00:38:40.326612 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:40 crc kubenswrapper[4888]: E1124 00:38:40.619942 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:41 crc kubenswrapper[4888]: I1124 00:38:41.226930 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:41 crc kubenswrapper[4888]: I1124 00:38:41.864890 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxzqq" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="registry-server" containerID="cri-o://99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46" gracePeriod=2 Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.342215 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.440963 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcmp8\" (UniqueName: \"kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8\") pod \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.441063 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities\") pod \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.441182 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content\") pod \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\" (UID: \"391ff92d-335b-4708-b8ca-b351cf6fa5ef\") " Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.442158 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities" (OuterVolumeSpecName: "utilities") pod "391ff92d-335b-4708-b8ca-b351cf6fa5ef" (UID: "391ff92d-335b-4708-b8ca-b351cf6fa5ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.449246 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8" (OuterVolumeSpecName: "kube-api-access-gcmp8") pod "391ff92d-335b-4708-b8ca-b351cf6fa5ef" (UID: "391ff92d-335b-4708-b8ca-b351cf6fa5ef"). InnerVolumeSpecName "kube-api-access-gcmp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.543073 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.543376 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcmp8\" (UniqueName: \"kubernetes.io/projected/391ff92d-335b-4708-b8ca-b351cf6fa5ef-kube-api-access-gcmp8\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.560487 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "391ff92d-335b-4708-b8ca-b351cf6fa5ef" (UID: "391ff92d-335b-4708-b8ca-b351cf6fa5ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.645524 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391ff92d-335b-4708-b8ca-b351cf6fa5ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.715578 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-mncvc"] Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716140 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="extract" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716212 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="extract" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716249 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="registry-server" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716265 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="registry-server" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716285 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="extract-utilities" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716300 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="extract-utilities" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716322 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="util" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716338 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="util" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716359 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="extract-content" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716373 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="extract-content" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.716397 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="pull" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716410 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="pull" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716635 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e749cc09-5587-4044-878a-c3adbd64909a" containerName="extract" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.716684 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerName="registry-server" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.717628 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.720508 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.728531 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-mncvc"] Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.729451 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4hrfx" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.729606 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.848530 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv85r\" (UniqueName: \"kubernetes.io/projected/f1648233-9c9f-4a7a-8acc-35d91c1a0b04-kube-api-access-zv85r\") pod \"nmstate-operator-557fdffb88-mncvc\" (UID: \"f1648233-9c9f-4a7a-8acc-35d91c1a0b04\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.875658 4888 generic.go:334] "Generic (PLEG): container finished" podID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" containerID="99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46" exitCode=0 Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.875706 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerDied","Data":"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46"} Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.875767 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxzqq" event={"ID":"391ff92d-335b-4708-b8ca-b351cf6fa5ef","Type":"ContainerDied","Data":"703b687ed031a94ce23687d445a0c76b9095bed2d3d05c83b2c926829e230fc6"} Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.875774 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxzqq" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.875796 4888 scope.go:117] "RemoveContainer" containerID="99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.894069 4888 scope.go:117] "RemoveContainer" containerID="b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.915039 4888 scope.go:117] "RemoveContainer" containerID="b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.959764 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv85r\" (UniqueName: \"kubernetes.io/projected/f1648233-9c9f-4a7a-8acc-35d91c1a0b04-kube-api-access-zv85r\") pod \"nmstate-operator-557fdffb88-mncvc\" (UID: \"f1648233-9c9f-4a7a-8acc-35d91c1a0b04\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.962350 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.969330 4888 scope.go:117] "RemoveContainer" containerID="99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.970206 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46\": container with ID starting with 99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46 not found: ID does not exist" containerID="99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.970244 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46"} err="failed to get container status \"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46\": rpc error: code = NotFound desc = could not find container \"99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46\": container with ID starting with 99c0750217b6326265675f9c3bb09e40981ef259f37991ea0e52f1d2bf066f46 not found: ID does not exist" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.970277 4888 scope.go:117] "RemoveContainer" containerID="b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.978252 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxzqq"] Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.981006 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be\": container with ID starting with b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be not found: ID does not exist" containerID="b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.981106 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be"} err="failed to get container status \"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be\": rpc error: code = NotFound desc = could not find container \"b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be\": container with ID starting with b511f538894a27178cbb13a3f5fa8a911a0509eb212f7249c40dfd1a719f22be not found: ID does not exist" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.981157 4888 scope.go:117] "RemoveContainer" containerID="b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec" Nov 24 00:38:42 crc kubenswrapper[4888]: E1124 00:38:42.982558 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec\": container with ID starting with b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec not found: ID does not exist" containerID="b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.982631 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec"} err="failed to get container status \"b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec\": rpc error: code = NotFound desc = could not find container \"b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec\": container with ID starting with b18ba2ee200e4d5e7c09b8194e53b7f1b7aa02c6075c319b3a8b5d834c4a3fec not found: ID does not exist" Nov 24 00:38:42 crc kubenswrapper[4888]: I1124 00:38:42.999023 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv85r\" (UniqueName: \"kubernetes.io/projected/f1648233-9c9f-4a7a-8acc-35d91c1a0b04-kube-api-access-zv85r\") pod \"nmstate-operator-557fdffb88-mncvc\" (UID: \"f1648233-9c9f-4a7a-8acc-35d91c1a0b04\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" Nov 24 00:38:43 crc kubenswrapper[4888]: I1124 00:38:43.050523 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" Nov 24 00:38:43 crc kubenswrapper[4888]: I1124 00:38:43.513064 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-mncvc"] Nov 24 00:38:43 crc kubenswrapper[4888]: W1124 00:38:43.517876 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1648233_9c9f_4a7a_8acc_35d91c1a0b04.slice/crio-37600f4f357405a848b866b6f7c4cb6c29742550a0ee7f58380a1451569db21a WatchSource:0}: Error finding container 37600f4f357405a848b866b6f7c4cb6c29742550a0ee7f58380a1451569db21a: Status 404 returned error can't find the container with id 37600f4f357405a848b866b6f7c4cb6c29742550a0ee7f58380a1451569db21a Nov 24 00:38:43 crc kubenswrapper[4888]: I1124 00:38:43.883411 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" event={"ID":"f1648233-9c9f-4a7a-8acc-35d91c1a0b04","Type":"ContainerStarted","Data":"37600f4f357405a848b866b6f7c4cb6c29742550a0ee7f58380a1451569db21a"} Nov 24 00:38:44 crc kubenswrapper[4888]: I1124 00:38:44.252976 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391ff92d-335b-4708-b8ca-b351cf6fa5ef" path="/var/lib/kubelet/pods/391ff92d-335b-4708-b8ca-b351cf6fa5ef/volumes" Nov 24 00:38:46 crc kubenswrapper[4888]: I1124 00:38:46.915469 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" event={"ID":"f1648233-9c9f-4a7a-8acc-35d91c1a0b04","Type":"ContainerStarted","Data":"950e9b813310286e9be8cddaaa86fc9ba86d0a7cfdcc9087840b936c4e3f3ed3"} Nov 24 00:38:46 crc kubenswrapper[4888]: I1124 00:38:46.956756 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-mncvc" podStartSLOduration=2.598788619 podStartE2EDuration="4.956728956s" podCreationTimestamp="2025-11-24 00:38:42 +0000 UTC" firstStartedPulling="2025-11-24 00:38:43.520366099 +0000 UTC m=+826.103050143" lastFinishedPulling="2025-11-24 00:38:45.878306426 +0000 UTC m=+828.460990480" observedRunningTime="2025-11-24 00:38:46.949136423 +0000 UTC m=+829.531820477" watchObservedRunningTime="2025-11-24 00:38:46.956728956 +0000 UTC m=+829.539413010" Nov 24 00:38:50 crc kubenswrapper[4888]: E1124 00:38:50.794565 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7312cd78_9bf2_4089_861f_37062cf2c728.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f172d7_c628_459d_9d8f_b664a18c08c5.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.686739 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.688337 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.691776 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zqlnk" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.701494 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.705265 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.706264 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.708420 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.715837 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fdwl9"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.716931 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.746131 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.746220 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2737508c-e47e-4ae4-b412-b665096fbf78-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.746715 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-ovs-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.746901 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5twcc\" (UniqueName: \"kubernetes.io/projected/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-kube-api-access-5twcc\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.747061 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-nmstate-lock\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.749423 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-dbus-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.750961 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96w8d\" (UniqueName: \"kubernetes.io/projected/f9bc22c3-35a5-4afc-90fe-67b9c086b46a-kube-api-access-96w8d\") pod \"nmstate-metrics-5dcf9c57c5-kgwmb\" (UID: \"f9bc22c3-35a5-4afc-90fe-67b9c086b46a\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.751345 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6jvp\" (UniqueName: \"kubernetes.io/projected/2737508c-e47e-4ae4-b412-b665096fbf78-kube-api-access-h6jvp\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.844170 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.845216 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.851385 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.851758 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.851769 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vxd6f" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852715 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96w8d\" (UniqueName: \"kubernetes.io/projected/f9bc22c3-35a5-4afc-90fe-67b9c086b46a-kube-api-access-96w8d\") pod \"nmstate-metrics-5dcf9c57c5-kgwmb\" (UID: \"f9bc22c3-35a5-4afc-90fe-67b9c086b46a\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852781 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6jvp\" (UniqueName: \"kubernetes.io/projected/2737508c-e47e-4ae4-b412-b665096fbf78-kube-api-access-h6jvp\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852863 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2737508c-e47e-4ae4-b412-b665096fbf78-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852892 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-ovs-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852920 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d88df599-7094-4325-afeb-d4188a9b948c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852954 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5twcc\" (UniqueName: \"kubernetes.io/projected/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-kube-api-access-5twcc\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.852986 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-nmstate-lock\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853027 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d88df599-7094-4325-afeb-d4188a9b948c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853053 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-dbus-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853089 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2psb8\" (UniqueName: \"kubernetes.io/projected/d88df599-7094-4325-afeb-d4188a9b948c-kube-api-access-2psb8\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853197 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-ovs-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853502 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-nmstate-lock\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.853832 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-dbus-socket\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.870020 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2737508c-e47e-4ae4-b412-b665096fbf78-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.872393 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2"] Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.879791 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6jvp\" (UniqueName: \"kubernetes.io/projected/2737508c-e47e-4ae4-b412-b665096fbf78-kube-api-access-h6jvp\") pod \"nmstate-webhook-6b89b748d8-fhrr5\" (UID: \"2737508c-e47e-4ae4-b412-b665096fbf78\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.880145 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5twcc\" (UniqueName: \"kubernetes.io/projected/00daa12e-54c4-46e9-835c-0a2d2c9ad90e-kube-api-access-5twcc\") pod \"nmstate-handler-fdwl9\" (UID: \"00daa12e-54c4-46e9-835c-0a2d2c9ad90e\") " pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.882739 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96w8d\" (UniqueName: \"kubernetes.io/projected/f9bc22c3-35a5-4afc-90fe-67b9c086b46a-kube-api-access-96w8d\") pod \"nmstate-metrics-5dcf9c57c5-kgwmb\" (UID: \"f9bc22c3-35a5-4afc-90fe-67b9c086b46a\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.955577 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d88df599-7094-4325-afeb-d4188a9b948c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.955929 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d88df599-7094-4325-afeb-d4188a9b948c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.955972 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2psb8\" (UniqueName: \"kubernetes.io/projected/d88df599-7094-4325-afeb-d4188a9b948c-kube-api-access-2psb8\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.956707 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d88df599-7094-4325-afeb-d4188a9b948c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.962509 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d88df599-7094-4325-afeb-d4188a9b948c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:52 crc kubenswrapper[4888]: I1124 00:38:52.978727 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2psb8\" (UniqueName: \"kubernetes.io/projected/d88df599-7094-4325-afeb-d4188a9b948c-kube-api-access-2psb8\") pod \"nmstate-console-plugin-5874bd7bc5-vzwm2\" (UID: \"d88df599-7094-4325-afeb-d4188a9b948c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.004910 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.021686 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.037153 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.078941 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.083988 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: W1124 00:38:53.125202 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00daa12e_54c4_46e9_835c_0a2d2c9ad90e.slice/crio-2c5945aa9cfaaa669a2589af06c9bbd2f4582f193037efdac0d75fec0b1b75c1 WatchSource:0}: Error finding container 2c5945aa9cfaaa669a2589af06c9bbd2f4582f193037efdac0d75fec0b1b75c1: Status 404 returned error can't find the container with id 2c5945aa9cfaaa669a2589af06c9bbd2f4582f193037efdac0d75fec0b1b75c1 Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.159640 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160843 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160916 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160934 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160959 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160977 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.160998 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.161043 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.233694 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.262999 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263066 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263088 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263109 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263168 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263227 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.263274 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.264504 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.265198 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.266444 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.266546 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.269609 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.269898 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.291680 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7\") pod \"console-597bdc9c68-fqxrz\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.339747 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb"] Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.402308 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.618083 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5"] Nov 24 00:38:53 crc kubenswrapper[4888]: W1124 00:38:53.627178 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2737508c_e47e_4ae4_b412_b665096fbf78.slice/crio-1bc5f08e02c677af47bc42161291eae0f9744749e532fa580bb4396a72354a3f WatchSource:0}: Error finding container 1bc5f08e02c677af47bc42161291eae0f9744749e532fa580bb4396a72354a3f: Status 404 returned error can't find the container with id 1bc5f08e02c677af47bc42161291eae0f9744749e532fa580bb4396a72354a3f Nov 24 00:38:53 crc kubenswrapper[4888]: I1124 00:38:53.783460 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2"] Nov 24 00:38:53 crc kubenswrapper[4888]: W1124 00:38:53.787252 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd88df599_7094_4325_afeb_d4188a9b948c.slice/crio-ead768c395272ac6b2a76e81c72f1376d5f6442848a3c256b2b319653800eefd WatchSource:0}: Error finding container ead768c395272ac6b2a76e81c72f1376d5f6442848a3c256b2b319653800eefd: Status 404 returned error can't find the container with id ead768c395272ac6b2a76e81c72f1376d5f6442848a3c256b2b319653800eefd Nov 24 00:38:54 crc kubenswrapper[4888]: I1124 00:38:54.005211 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:38:54 crc kubenswrapper[4888]: I1124 00:38:54.006046 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" event={"ID":"2737508c-e47e-4ae4-b412-b665096fbf78","Type":"ContainerStarted","Data":"1bc5f08e02c677af47bc42161291eae0f9744749e532fa580bb4396a72354a3f"} Nov 24 00:38:54 crc kubenswrapper[4888]: I1124 00:38:54.025285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fdwl9" event={"ID":"00daa12e-54c4-46e9-835c-0a2d2c9ad90e","Type":"ContainerStarted","Data":"2c5945aa9cfaaa669a2589af06c9bbd2f4582f193037efdac0d75fec0b1b75c1"} Nov 24 00:38:54 crc kubenswrapper[4888]: I1124 00:38:54.045946 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" event={"ID":"d88df599-7094-4325-afeb-d4188a9b948c","Type":"ContainerStarted","Data":"ead768c395272ac6b2a76e81c72f1376d5f6442848a3c256b2b319653800eefd"} Nov 24 00:38:54 crc kubenswrapper[4888]: I1124 00:38:54.061984 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" event={"ID":"f9bc22c3-35a5-4afc-90fe-67b9c086b46a","Type":"ContainerStarted","Data":"47bf6b5bec621d9aa4567612a25085ac95a685f4826594849c48942464c69da8"} Nov 24 00:38:55 crc kubenswrapper[4888]: I1124 00:38:55.074166 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-597bdc9c68-fqxrz" event={"ID":"d41a16d1-0335-4e9e-b64d-f21337fcbe82","Type":"ContainerStarted","Data":"d1e6ca6ed1d93ac947861b41f240dba5684f9128dd4aa37488943c91b6f6e07c"} Nov 24 00:38:55 crc kubenswrapper[4888]: I1124 00:38:55.075024 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-597bdc9c68-fqxrz" event={"ID":"d41a16d1-0335-4e9e-b64d-f21337fcbe82","Type":"ContainerStarted","Data":"8b0f7c68853638d22eeb419c2d643928ca73ed792bf3f66172d1566c64b88459"} Nov 24 00:38:55 crc kubenswrapper[4888]: I1124 00:38:55.106972 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-597bdc9c68-fqxrz" podStartSLOduration=2.10694253 podStartE2EDuration="2.10694253s" podCreationTimestamp="2025-11-24 00:38:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:38:55.0955488 +0000 UTC m=+837.678232864" watchObservedRunningTime="2025-11-24 00:38:55.10694253 +0000 UTC m=+837.689626584" Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.093011 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" event={"ID":"2737508c-e47e-4ae4-b412-b665096fbf78","Type":"ContainerStarted","Data":"5373fa3904a462521100c3c6b3098ba6c7a961f59c0aa4dc457b1678620f00ee"} Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.093279 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.094917 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fdwl9" event={"ID":"00daa12e-54c4-46e9-835c-0a2d2c9ad90e","Type":"ContainerStarted","Data":"3ed1b26100d63b487833226f7c5d73630cf0e2e1c808254a733904d3708faab5"} Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.095035 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.096843 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" event={"ID":"d88df599-7094-4325-afeb-d4188a9b948c","Type":"ContainerStarted","Data":"1d4f6c0a40ac4c5c8c5314c7519c1f5259087331a440fa9eb23a4b2c9240c1eb"} Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.098114 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" event={"ID":"f9bc22c3-35a5-4afc-90fe-67b9c086b46a","Type":"ContainerStarted","Data":"60eeac7a0876576309630dcc265c78c8850c88303b5289da912204509e3862ee"} Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.113741 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" podStartSLOduration=2.045342183 podStartE2EDuration="5.113726634s" podCreationTimestamp="2025-11-24 00:38:52 +0000 UTC" firstStartedPulling="2025-11-24 00:38:53.629019359 +0000 UTC m=+836.211703393" lastFinishedPulling="2025-11-24 00:38:56.69740376 +0000 UTC m=+839.280087844" observedRunningTime="2025-11-24 00:38:57.112385926 +0000 UTC m=+839.695069990" watchObservedRunningTime="2025-11-24 00:38:57.113726634 +0000 UTC m=+839.696410678" Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.131095 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-vzwm2" podStartSLOduration=2.247250054 podStartE2EDuration="5.131067741s" podCreationTimestamp="2025-11-24 00:38:52 +0000 UTC" firstStartedPulling="2025-11-24 00:38:53.789549418 +0000 UTC m=+836.372233472" lastFinishedPulling="2025-11-24 00:38:56.673367085 +0000 UTC m=+839.256051159" observedRunningTime="2025-11-24 00:38:57.127439349 +0000 UTC m=+839.710123393" watchObservedRunningTime="2025-11-24 00:38:57.131067741 +0000 UTC m=+839.713751785" Nov 24 00:38:57 crc kubenswrapper[4888]: I1124 00:38:57.146781 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fdwl9" podStartSLOduration=1.593122642 podStartE2EDuration="5.146760122s" podCreationTimestamp="2025-11-24 00:38:52 +0000 UTC" firstStartedPulling="2025-11-24 00:38:53.143803711 +0000 UTC m=+835.726487755" lastFinishedPulling="2025-11-24 00:38:56.697441151 +0000 UTC m=+839.280125235" observedRunningTime="2025-11-24 00:38:57.144438476 +0000 UTC m=+839.727122530" watchObservedRunningTime="2025-11-24 00:38:57.146760122 +0000 UTC m=+839.729444166" Nov 24 00:39:00 crc kubenswrapper[4888]: I1124 00:39:00.166787 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" event={"ID":"f9bc22c3-35a5-4afc-90fe-67b9c086b46a","Type":"ContainerStarted","Data":"dce55def290b48eabfc3e39ebb408937b6323635ecf4bb0c61a1ac562551b3bd"} Nov 24 00:39:00 crc kubenswrapper[4888]: I1124 00:39:00.205482 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kgwmb" podStartSLOduration=2.060986621 podStartE2EDuration="8.205409159s" podCreationTimestamp="2025-11-24 00:38:52 +0000 UTC" firstStartedPulling="2025-11-24 00:38:53.35490434 +0000 UTC m=+835.937588394" lastFinishedPulling="2025-11-24 00:38:59.499326888 +0000 UTC m=+842.082010932" observedRunningTime="2025-11-24 00:39:00.193274388 +0000 UTC m=+842.775958472" watchObservedRunningTime="2025-11-24 00:39:00.205409159 +0000 UTC m=+842.788093233" Nov 24 00:39:03 crc kubenswrapper[4888]: I1124 00:39:03.071794 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fdwl9" Nov 24 00:39:03 crc kubenswrapper[4888]: I1124 00:39:03.404070 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:39:03 crc kubenswrapper[4888]: I1124 00:39:03.404195 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:39:03 crc kubenswrapper[4888]: I1124 00:39:03.412849 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:39:04 crc kubenswrapper[4888]: I1124 00:39:04.209765 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:39:04 crc kubenswrapper[4888]: I1124 00:39:04.283508 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:39:13 crc kubenswrapper[4888]: I1124 00:39:13.031623 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fhrr5" Nov 24 00:39:23 crc kubenswrapper[4888]: I1124 00:39:23.591958 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:39:23 crc kubenswrapper[4888]: I1124 00:39:23.592636 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.359143 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-jdqr8" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" containerName="console" containerID="cri-o://b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971" gracePeriod=15 Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.825531 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jdqr8_e594a1de-76b4-4c46-8244-8188ece0c088/console/0.log" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.825887 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.911743 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.911828 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.911908 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.911994 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.912071 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.912103 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcwnf\" (UniqueName: \"kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.912140 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert\") pod \"e594a1de-76b4-4c46-8244-8188ece0c088\" (UID: \"e594a1de-76b4-4c46-8244-8188ece0c088\") " Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.913103 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca" (OuterVolumeSpecName: "service-ca") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.913526 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.913531 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config" (OuterVolumeSpecName: "console-config") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.913704 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.914511 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.919612 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.919775 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:39:29 crc kubenswrapper[4888]: I1124 00:39:29.919795 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf" (OuterVolumeSpecName: "kube-api-access-jcwnf") pod "e594a1de-76b4-4c46-8244-8188ece0c088" (UID: "e594a1de-76b4-4c46-8244-8188ece0c088"). InnerVolumeSpecName "kube-api-access-jcwnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016624 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcwnf\" (UniqueName: \"kubernetes.io/projected/e594a1de-76b4-4c46-8244-8188ece0c088-kube-api-access-jcwnf\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016658 4888 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016667 4888 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e594a1de-76b4-4c46-8244-8188ece0c088-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016677 4888 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016686 4888 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.016694 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e594a1de-76b4-4c46-8244-8188ece0c088-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477504 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jdqr8_e594a1de-76b4-4c46-8244-8188ece0c088/console/0.log" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477628 4888 generic.go:334] "Generic (PLEG): container finished" podID="e594a1de-76b4-4c46-8244-8188ece0c088" containerID="b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971" exitCode=2 Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477699 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jdqr8" event={"ID":"e594a1de-76b4-4c46-8244-8188ece0c088","Type":"ContainerDied","Data":"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971"} Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477774 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jdqr8" event={"ID":"e594a1de-76b4-4c46-8244-8188ece0c088","Type":"ContainerDied","Data":"cffbe386c1fde27a96e1e870208c1e9b644617a5485906e999245fedf1c660cf"} Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477846 4888 scope.go:117] "RemoveContainer" containerID="b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.477864 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jdqr8" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.504274 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.509910 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-jdqr8"] Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.512903 4888 scope.go:117] "RemoveContainer" containerID="b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971" Nov 24 00:39:30 crc kubenswrapper[4888]: E1124 00:39:30.513519 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971\": container with ID starting with b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971 not found: ID does not exist" containerID="b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971" Nov 24 00:39:30 crc kubenswrapper[4888]: I1124 00:39:30.513569 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971"} err="failed to get container status \"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971\": rpc error: code = NotFound desc = could not find container \"b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971\": container with ID starting with b67c1a53a851a421fa54e908f27b73ca90c0dfbbd6081b07411a65f9a1db6971 not found: ID does not exist" Nov 24 00:39:32 crc kubenswrapper[4888]: I1124 00:39:32.254228 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" path="/var/lib/kubelet/pods/e594a1de-76b4-4c46-8244-8188ece0c088/volumes" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.048440 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5"] Nov 24 00:39:35 crc kubenswrapper[4888]: E1124 00:39:35.049281 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" containerName="console" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.049303 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" containerName="console" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.049630 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e594a1de-76b4-4c46-8244-8188ece0c088" containerName="console" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.051645 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.054453 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.077433 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5"] Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.112406 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.112490 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.112598 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86c22\" (UniqueName: \"kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.214880 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.214949 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.214997 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86c22\" (UniqueName: \"kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.215472 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.215798 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.243343 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86c22\" (UniqueName: \"kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.387456 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:35 crc kubenswrapper[4888]: I1124 00:39:35.891546 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5"] Nov 24 00:39:36 crc kubenswrapper[4888]: I1124 00:39:36.527313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" event={"ID":"8210c0ef-1d11-4f4f-b05d-b584490ecd29","Type":"ContainerStarted","Data":"1c752b879de943b53c3de17359e37a735a096412b5f5d2dcb9f40b9c84ff649c"} Nov 24 00:39:37 crc kubenswrapper[4888]: I1124 00:39:37.536528 4888 generic.go:334] "Generic (PLEG): container finished" podID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerID="16996bd48460bffd8222fa4dc81488260fe40abb03f7987cd2082c2a32a302b1" exitCode=0 Nov 24 00:39:37 crc kubenswrapper[4888]: I1124 00:39:37.536582 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" event={"ID":"8210c0ef-1d11-4f4f-b05d-b584490ecd29","Type":"ContainerDied","Data":"16996bd48460bffd8222fa4dc81488260fe40abb03f7987cd2082c2a32a302b1"} Nov 24 00:39:37 crc kubenswrapper[4888]: I1124 00:39:37.538941 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:39:39 crc kubenswrapper[4888]: I1124 00:39:39.561425 4888 generic.go:334] "Generic (PLEG): container finished" podID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerID="8615a75cadf359936a3cb8a2404ee145095a6232b9c35ce9418ca7ee273725e7" exitCode=0 Nov 24 00:39:39 crc kubenswrapper[4888]: I1124 00:39:39.561583 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" event={"ID":"8210c0ef-1d11-4f4f-b05d-b584490ecd29","Type":"ContainerDied","Data":"8615a75cadf359936a3cb8a2404ee145095a6232b9c35ce9418ca7ee273725e7"} Nov 24 00:39:40 crc kubenswrapper[4888]: I1124 00:39:40.575740 4888 generic.go:334] "Generic (PLEG): container finished" podID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerID="defa2e3eaa91bee87b3fdbf9886db7cf00cf1c40a2ec0548694bae85fb1b7453" exitCode=0 Nov 24 00:39:40 crc kubenswrapper[4888]: I1124 00:39:40.575943 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" event={"ID":"8210c0ef-1d11-4f4f-b05d-b584490ecd29","Type":"ContainerDied","Data":"defa2e3eaa91bee87b3fdbf9886db7cf00cf1c40a2ec0548694bae85fb1b7453"} Nov 24 00:39:41 crc kubenswrapper[4888]: I1124 00:39:41.894707 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.056702 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util\") pod \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.056875 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle\") pod \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.056987 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86c22\" (UniqueName: \"kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22\") pod \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\" (UID: \"8210c0ef-1d11-4f4f-b05d-b584490ecd29\") " Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.058969 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle" (OuterVolumeSpecName: "bundle") pod "8210c0ef-1d11-4f4f-b05d-b584490ecd29" (UID: "8210c0ef-1d11-4f4f-b05d-b584490ecd29"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.068061 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22" (OuterVolumeSpecName: "kube-api-access-86c22") pod "8210c0ef-1d11-4f4f-b05d-b584490ecd29" (UID: "8210c0ef-1d11-4f4f-b05d-b584490ecd29"). InnerVolumeSpecName "kube-api-access-86c22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.159225 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86c22\" (UniqueName: \"kubernetes.io/projected/8210c0ef-1d11-4f4f-b05d-b584490ecd29-kube-api-access-86c22\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.159279 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.416908 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util" (OuterVolumeSpecName: "util") pod "8210c0ef-1d11-4f4f-b05d-b584490ecd29" (UID: "8210c0ef-1d11-4f4f-b05d-b584490ecd29"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.464353 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8210c0ef-1d11-4f4f-b05d-b584490ecd29-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.594294 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" event={"ID":"8210c0ef-1d11-4f4f-b05d-b584490ecd29","Type":"ContainerDied","Data":"1c752b879de943b53c3de17359e37a735a096412b5f5d2dcb9f40b9c84ff649c"} Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.594363 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c752b879de943b53c3de17359e37a735a096412b5f5d2dcb9f40b9c84ff649c" Nov 24 00:39:42 crc kubenswrapper[4888]: I1124 00:39:42.594399 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.943528 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm"] Nov 24 00:39:52 crc kubenswrapper[4888]: E1124 00:39:52.944351 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="extract" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.944381 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="extract" Nov 24 00:39:52 crc kubenswrapper[4888]: E1124 00:39:52.944401 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="pull" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.944406 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="pull" Nov 24 00:39:52 crc kubenswrapper[4888]: E1124 00:39:52.944425 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="util" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.944432 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="util" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.944550 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8210c0ef-1d11-4f4f-b05d-b584490ecd29" containerName="extract" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.945183 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.948394 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-6l8lj" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.948700 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.949064 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.949380 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.963723 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 00:39:52 crc kubenswrapper[4888]: I1124 00:39:52.973716 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm"] Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.045003 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.045083 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwqx5\" (UniqueName: \"kubernetes.io/projected/d8651fd8-33fc-4d15-b478-78921c04fef1-kube-api-access-hwqx5\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.045197 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-webhook-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.146600 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-webhook-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.146672 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.146715 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwqx5\" (UniqueName: \"kubernetes.io/projected/d8651fd8-33fc-4d15-b478-78921c04fef1-kube-api-access-hwqx5\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.155405 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-webhook-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.165520 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8651fd8-33fc-4d15-b478-78921c04fef1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.173637 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwqx5\" (UniqueName: \"kubernetes.io/projected/d8651fd8-33fc-4d15-b478-78921c04fef1-kube-api-access-hwqx5\") pod \"metallb-operator-controller-manager-6f89595f9b-rs4gm\" (UID: \"d8651fd8-33fc-4d15-b478-78921c04fef1\") " pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.261303 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.411996 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn"] Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.413344 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.418322 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.418755 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.419307 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-7kg6k" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.479369 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn"] Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.554803 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-apiservice-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.554928 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-webhook-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.554952 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shswv\" (UniqueName: \"kubernetes.io/projected/007267a1-8af0-47c2-84e1-8fa533e3f5c7-kube-api-access-shswv\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.591899 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.592282 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.657165 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-webhook-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.657219 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shswv\" (UniqueName: \"kubernetes.io/projected/007267a1-8af0-47c2-84e1-8fa533e3f5c7-kube-api-access-shswv\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.657297 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-apiservice-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.665556 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-apiservice-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.678362 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007267a1-8af0-47c2-84e1-8fa533e3f5c7-webhook-cert\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.685787 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shswv\" (UniqueName: \"kubernetes.io/projected/007267a1-8af0-47c2-84e1-8fa533e3f5c7-kube-api-access-shswv\") pod \"metallb-operator-webhook-server-69f6b76c88-dh8sn\" (UID: \"007267a1-8af0-47c2-84e1-8fa533e3f5c7\") " pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.732786 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.870230 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm"] Nov 24 00:39:53 crc kubenswrapper[4888]: I1124 00:39:53.975183 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn"] Nov 24 00:39:53 crc kubenswrapper[4888]: W1124 00:39:53.983196 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod007267a1_8af0_47c2_84e1_8fa533e3f5c7.slice/crio-6c52a7da24c0b8a7bc0915cbf9a04567948e284c89ec96b3a41b4109723a3f29 WatchSource:0}: Error finding container 6c52a7da24c0b8a7bc0915cbf9a04567948e284c89ec96b3a41b4109723a3f29: Status 404 returned error can't find the container with id 6c52a7da24c0b8a7bc0915cbf9a04567948e284c89ec96b3a41b4109723a3f29 Nov 24 00:39:54 crc kubenswrapper[4888]: I1124 00:39:54.698895 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" event={"ID":"007267a1-8af0-47c2-84e1-8fa533e3f5c7","Type":"ContainerStarted","Data":"6c52a7da24c0b8a7bc0915cbf9a04567948e284c89ec96b3a41b4109723a3f29"} Nov 24 00:39:54 crc kubenswrapper[4888]: I1124 00:39:54.700995 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" event={"ID":"d8651fd8-33fc-4d15-b478-78921c04fef1","Type":"ContainerStarted","Data":"5d7c3c8fa368f016f0a5401cdf6de9585196dec95efe18ab5e8e6ec0e7a47770"} Nov 24 00:39:57 crc kubenswrapper[4888]: I1124 00:39:57.735212 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" event={"ID":"d8651fd8-33fc-4d15-b478-78921c04fef1","Type":"ContainerStarted","Data":"dc452624c8f1a3fc471fbbf047d1f849b030aa6ddab8c36526b62bedd84f34c8"} Nov 24 00:39:57 crc kubenswrapper[4888]: I1124 00:39:57.735907 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:39:57 crc kubenswrapper[4888]: I1124 00:39:57.760239 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" podStartSLOduration=2.36577772 podStartE2EDuration="5.76022066s" podCreationTimestamp="2025-11-24 00:39:52 +0000 UTC" firstStartedPulling="2025-11-24 00:39:53.880521761 +0000 UTC m=+896.463205805" lastFinishedPulling="2025-11-24 00:39:57.274964701 +0000 UTC m=+899.857648745" observedRunningTime="2025-11-24 00:39:57.758621105 +0000 UTC m=+900.341305149" watchObservedRunningTime="2025-11-24 00:39:57.76022066 +0000 UTC m=+900.342904694" Nov 24 00:39:59 crc kubenswrapper[4888]: I1124 00:39:59.756970 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" event={"ID":"007267a1-8af0-47c2-84e1-8fa533e3f5c7","Type":"ContainerStarted","Data":"d7ee5652b3c8e6a6d4891147aa13d7457a4f1cc11bdd9775aa17c7c8022e6e7b"} Nov 24 00:39:59 crc kubenswrapper[4888]: I1124 00:39:59.757434 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:39:59 crc kubenswrapper[4888]: I1124 00:39:59.778499 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" podStartSLOduration=1.337368013 podStartE2EDuration="6.778475127s" podCreationTimestamp="2025-11-24 00:39:53 +0000 UTC" firstStartedPulling="2025-11-24 00:39:53.986165069 +0000 UTC m=+896.568849113" lastFinishedPulling="2025-11-24 00:39:59.427272183 +0000 UTC m=+902.009956227" observedRunningTime="2025-11-24 00:39:59.776194993 +0000 UTC m=+902.358879077" watchObservedRunningTime="2025-11-24 00:39:59.778475127 +0000 UTC m=+902.361159171" Nov 24 00:40:13 crc kubenswrapper[4888]: I1124 00:40:13.738630 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-69f6b76c88-dh8sn" Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.591684 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.592761 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.592907 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.594114 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.594379 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778" gracePeriod=600 Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.981400 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778" exitCode=0 Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.981886 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778"} Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.981933 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5"} Nov 24 00:40:23 crc kubenswrapper[4888]: I1124 00:40:23.981963 4888 scope.go:117] "RemoveContainer" containerID="e7751efa01bd9b40cbb0f03161ac79b109eb34ebe03b2ce801b875ba7b0165d4" Nov 24 00:40:33 crc kubenswrapper[4888]: I1124 00:40:33.267077 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6f89595f9b-rs4gm" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.200501 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bdrm6"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.205070 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.227061 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.227311 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6s9gc" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.227626 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.262464 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.264867 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.264992 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.273296 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.307914 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5b721b4f-a010-471e-a766-159936fc9379-frr-startup\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308031 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-metrics\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308151 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-conf\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308249 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jlvl\" (UniqueName: \"kubernetes.io/projected/5b721b4f-a010-471e-a766-159936fc9379-kube-api-access-2jlvl\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308440 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-reloader\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308503 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-sockets\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.308529 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.347917 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-2lxkv"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.349340 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.354863 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-psz9x" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.355204 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.355311 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.355435 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.358850 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-77rp4"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.360266 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.363581 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.364346 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-77rp4"] Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410085 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jlvl\" (UniqueName: \"kubernetes.io/projected/5b721b4f-a010-471e-a766-159936fc9379-kube-api-access-2jlvl\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410172 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43445b39-e7aa-4d6c-bc1a-760e001605b3-cert\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410227 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-reloader\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410253 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh6k9\" (UniqueName: \"kubernetes.io/projected/43445b39-e7aa-4d6c-bc1a-760e001605b3-kube-api-access-wh6k9\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410285 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-sockets\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410301 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410318 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5b721b4f-a010-471e-a766-159936fc9379-frr-startup\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410339 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-metrics\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.410380 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-conf\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: E1124 00:40:34.410916 4888 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 00:40:34 crc kubenswrapper[4888]: E1124 00:40:34.410985 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs podName:5b721b4f-a010-471e-a766-159936fc9379 nodeName:}" failed. No retries permitted until 2025-11-24 00:40:34.910966513 +0000 UTC m=+937.493650557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs") pod "frr-k8s-bdrm6" (UID: "5b721b4f-a010-471e-a766-159936fc9379") : secret "frr-k8s-certs-secret" not found Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.411143 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-reloader\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.411363 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-conf\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.411404 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-metrics\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.411555 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5b721b4f-a010-471e-a766-159936fc9379-frr-sockets\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.412422 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5b721b4f-a010-471e-a766-159936fc9379-frr-startup\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.444909 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jlvl\" (UniqueName: \"kubernetes.io/projected/5b721b4f-a010-471e-a766-159936fc9379-kube-api-access-2jlvl\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511410 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-metrics-certs\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511464 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511506 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9htv\" (UniqueName: \"kubernetes.io/projected/2b50d2f6-eedb-4817-ba29-4b6e908aae96-kube-api-access-q9htv\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511611 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-cert\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511657 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metrics-certs\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511683 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w46gj\" (UniqueName: \"kubernetes.io/projected/e710f559-91d6-47ce-bf7d-ee5717c276a8-kube-api-access-w46gj\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511728 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43445b39-e7aa-4d6c-bc1a-760e001605b3-cert\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511765 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh6k9\" (UniqueName: \"kubernetes.io/projected/43445b39-e7aa-4d6c-bc1a-760e001605b3-kube-api-access-wh6k9\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.511805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metallb-excludel2\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.520087 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43445b39-e7aa-4d6c-bc1a-760e001605b3-cert\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.535546 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh6k9\" (UniqueName: \"kubernetes.io/projected/43445b39-e7aa-4d6c-bc1a-760e001605b3-kube-api-access-wh6k9\") pod \"frr-k8s-webhook-server-6998585d5-zvfv2\" (UID: \"43445b39-e7aa-4d6c-bc1a-760e001605b3\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.611453 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614338 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9htv\" (UniqueName: \"kubernetes.io/projected/2b50d2f6-eedb-4817-ba29-4b6e908aae96-kube-api-access-q9htv\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614433 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-cert\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614455 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metrics-certs\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614479 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w46gj\" (UniqueName: \"kubernetes.io/projected/e710f559-91d6-47ce-bf7d-ee5717c276a8-kube-api-access-w46gj\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614549 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metallb-excludel2\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.614576 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-metrics-certs\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: E1124 00:40:34.614635 4888 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 00:40:34 crc kubenswrapper[4888]: E1124 00:40:34.614758 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist podName:2b50d2f6-eedb-4817-ba29-4b6e908aae96 nodeName:}" failed. No retries permitted until 2025-11-24 00:40:35.114728066 +0000 UTC m=+937.697412150 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist") pod "speaker-2lxkv" (UID: "2b50d2f6-eedb-4817-ba29-4b6e908aae96") : secret "metallb-memberlist" not found Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.616318 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metallb-excludel2\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.617326 4888 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.618373 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-metrics-certs\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.619357 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-metrics-certs\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.629557 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e710f559-91d6-47ce-bf7d-ee5717c276a8-cert\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.634277 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9htv\" (UniqueName: \"kubernetes.io/projected/2b50d2f6-eedb-4817-ba29-4b6e908aae96-kube-api-access-q9htv\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.636188 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w46gj\" (UniqueName: \"kubernetes.io/projected/e710f559-91d6-47ce-bf7d-ee5717c276a8-kube-api-access-w46gj\") pod \"controller-6c7b4b5f48-77rp4\" (UID: \"e710f559-91d6-47ce-bf7d-ee5717c276a8\") " pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.695235 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.920018 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:34 crc kubenswrapper[4888]: I1124 00:40:34.924327 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b721b4f-a010-471e-a766-159936fc9379-metrics-certs\") pod \"frr-k8s-bdrm6\" (UID: \"5b721b4f-a010-471e-a766-159936fc9379\") " pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:35 crc kubenswrapper[4888]: I1124 00:40:35.123783 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:35 crc kubenswrapper[4888]: E1124 00:40:35.124278 4888 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 00:40:35 crc kubenswrapper[4888]: E1124 00:40:35.124378 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist podName:2b50d2f6-eedb-4817-ba29-4b6e908aae96 nodeName:}" failed. No retries permitted until 2025-11-24 00:40:36.12435146 +0000 UTC m=+938.707035544 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist") pod "speaker-2lxkv" (UID: "2b50d2f6-eedb-4817-ba29-4b6e908aae96") : secret "metallb-memberlist" not found Nov 24 00:40:35 crc kubenswrapper[4888]: I1124 00:40:35.126666 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:35 crc kubenswrapper[4888]: I1124 00:40:35.177886 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-77rp4"] Nov 24 00:40:35 crc kubenswrapper[4888]: I1124 00:40:35.194787 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2"] Nov 24 00:40:35 crc kubenswrapper[4888]: W1124 00:40:35.202570 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43445b39_e7aa_4d6c_bc1a_760e001605b3.slice/crio-3eadf426693af25f8df8a6363cf5baa0ac4768de2dc318733eaef3a41c7c9254 WatchSource:0}: Error finding container 3eadf426693af25f8df8a6363cf5baa0ac4768de2dc318733eaef3a41c7c9254: Status 404 returned error can't find the container with id 3eadf426693af25f8df8a6363cf5baa0ac4768de2dc318733eaef3a41c7c9254 Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.105393 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-77rp4" event={"ID":"e710f559-91d6-47ce-bf7d-ee5717c276a8","Type":"ContainerStarted","Data":"ea6d6060983591d010afdcc8059a57475672c3f2874a3ef145f32dcf8daf5899"} Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.105734 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.105750 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-77rp4" event={"ID":"e710f559-91d6-47ce-bf7d-ee5717c276a8","Type":"ContainerStarted","Data":"4706031ec525cee1a1c5b43890c396054095981fddfefbe9c506929efba94018"} Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.105760 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-77rp4" event={"ID":"e710f559-91d6-47ce-bf7d-ee5717c276a8","Type":"ContainerStarted","Data":"c0c7177e4461da58a24d976dcf89868031d14277f069b62c54ce2912287cde5e"} Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.107074 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" event={"ID":"43445b39-e7aa-4d6c-bc1a-760e001605b3","Type":"ContainerStarted","Data":"3eadf426693af25f8df8a6363cf5baa0ac4768de2dc318733eaef3a41c7c9254"} Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.108560 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"4549524ef52ca90481bfcfc3e5799667f9e61dbad8a36e9ec642c89a93d7052e"} Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.139983 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.149122 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2b50d2f6-eedb-4817-ba29-4b6e908aae96-memberlist\") pod \"speaker-2lxkv\" (UID: \"2b50d2f6-eedb-4817-ba29-4b6e908aae96\") " pod="metallb-system/speaker-2lxkv" Nov 24 00:40:36 crc kubenswrapper[4888]: I1124 00:40:36.187759 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2lxkv" Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.119722 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2lxkv" event={"ID":"2b50d2f6-eedb-4817-ba29-4b6e908aae96","Type":"ContainerStarted","Data":"98f6ac26898d8115bc0bee5f80c2992bdaa131d3758defa7fdb6a57bc32c16d1"} Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.120188 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2lxkv" event={"ID":"2b50d2f6-eedb-4817-ba29-4b6e908aae96","Type":"ContainerStarted","Data":"848b5d8ec077e27c86971e8431d1cdd3879cfb140a8802f2f8af4af1bab56113"} Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.120201 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2lxkv" event={"ID":"2b50d2f6-eedb-4817-ba29-4b6e908aae96","Type":"ContainerStarted","Data":"9e86c1449373355b654bd8767fe992927d4142540cfb9e7c3ea528923b9ebf17"} Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.120400 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-2lxkv" Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.156189 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-77rp4" podStartSLOduration=3.156167307 podStartE2EDuration="3.156167307s" podCreationTimestamp="2025-11-24 00:40:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:40:36.127585788 +0000 UTC m=+938.710269842" watchObservedRunningTime="2025-11-24 00:40:37.156167307 +0000 UTC m=+939.738851351" Nov 24 00:40:37 crc kubenswrapper[4888]: I1124 00:40:37.157173 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-2lxkv" podStartSLOduration=3.157167005 podStartE2EDuration="3.157167005s" podCreationTimestamp="2025-11-24 00:40:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:40:37.149716306 +0000 UTC m=+939.732400350" watchObservedRunningTime="2025-11-24 00:40:37.157167005 +0000 UTC m=+939.739851049" Nov 24 00:40:43 crc kubenswrapper[4888]: I1124 00:40:43.179980 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" event={"ID":"43445b39-e7aa-4d6c-bc1a-760e001605b3","Type":"ContainerStarted","Data":"cb7acb91d1144f923276207881926fee9da2b9028bd3b08937b85446a825a3af"} Nov 24 00:40:43 crc kubenswrapper[4888]: I1124 00:40:43.181006 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:43 crc kubenswrapper[4888]: I1124 00:40:43.182387 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b721b4f-a010-471e-a766-159936fc9379" containerID="166cfb99f7f3b02c25c3e4dd2693ff739042b8de0f6c00b4eb8f0f9ab8e22011" exitCode=0 Nov 24 00:40:43 crc kubenswrapper[4888]: I1124 00:40:43.182456 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerDied","Data":"166cfb99f7f3b02c25c3e4dd2693ff739042b8de0f6c00b4eb8f0f9ab8e22011"} Nov 24 00:40:43 crc kubenswrapper[4888]: I1124 00:40:43.207887 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" podStartSLOduration=1.799368689 podStartE2EDuration="9.20785904s" podCreationTimestamp="2025-11-24 00:40:34 +0000 UTC" firstStartedPulling="2025-11-24 00:40:35.204273885 +0000 UTC m=+937.786957929" lastFinishedPulling="2025-11-24 00:40:42.612764246 +0000 UTC m=+945.195448280" observedRunningTime="2025-11-24 00:40:43.206859642 +0000 UTC m=+945.789543726" watchObservedRunningTime="2025-11-24 00:40:43.20785904 +0000 UTC m=+945.790543124" Nov 24 00:40:44 crc kubenswrapper[4888]: I1124 00:40:44.196993 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b721b4f-a010-471e-a766-159936fc9379" containerID="2394ac43919a0125164c1e89ac5e181d55299c5b2e1c708f13042489d8bb4b46" exitCode=0 Nov 24 00:40:44 crc kubenswrapper[4888]: I1124 00:40:44.197065 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerDied","Data":"2394ac43919a0125164c1e89ac5e181d55299c5b2e1c708f13042489d8bb4b46"} Nov 24 00:40:45 crc kubenswrapper[4888]: I1124 00:40:45.207793 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b721b4f-a010-471e-a766-159936fc9379" containerID="ef69f7e160da4b1367cd44fc86a7ad9be84a75ad822eec2528748367718fc1fb" exitCode=0 Nov 24 00:40:45 crc kubenswrapper[4888]: I1124 00:40:45.207865 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerDied","Data":"ef69f7e160da4b1367cd44fc86a7ad9be84a75ad822eec2528748367718fc1fb"} Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.193102 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-2lxkv" Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.231779 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"dc916d39dddc942d60af98bb1ef70af5f74cb13bb3ff9c66627d48aeed3c6be9"} Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.231860 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"fd48d7806ac9f792249ed04470bbf222cd3e62f3d9e727343c2c718b4b2ff257"} Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.231878 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"490b920be0496aafc7b1062681f604f819e599bee72f143dfd2b639a7cdb952f"} Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.231893 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"0d8b4c1ae340759a5cb3af306c7c2714f172fb762b68248c6ebc4ed21d69329f"} Nov 24 00:40:46 crc kubenswrapper[4888]: I1124 00:40:46.231911 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"9c9c49d4163fc54559e6f791d34b0b4b95c0a88cd38d7bb825ba5999216d43f0"} Nov 24 00:40:47 crc kubenswrapper[4888]: I1124 00:40:47.249999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bdrm6" event={"ID":"5b721b4f-a010-471e-a766-159936fc9379","Type":"ContainerStarted","Data":"dd7dff69755e708e94a25d0ea39bb4997ff2c91c6c7ca120deb6d1c0a098785b"} Nov 24 00:40:47 crc kubenswrapper[4888]: I1124 00:40:47.250558 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:47 crc kubenswrapper[4888]: I1124 00:40:47.288147 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bdrm6" podStartSLOduration=5.913921353 podStartE2EDuration="13.288114451s" podCreationTimestamp="2025-11-24 00:40:34 +0000 UTC" firstStartedPulling="2025-11-24 00:40:35.271699248 +0000 UTC m=+937.854383282" lastFinishedPulling="2025-11-24 00:40:42.645892296 +0000 UTC m=+945.228576380" observedRunningTime="2025-11-24 00:40:47.286220838 +0000 UTC m=+949.868904932" watchObservedRunningTime="2025-11-24 00:40:47.288114451 +0000 UTC m=+949.870798535" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.284553 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.287003 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.291639 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.294469 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.294559 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.294721 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xf467" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.443213 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-668ss\" (UniqueName: \"kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss\") pod \"openstack-operator-index-dzdn2\" (UID: \"9dda7a38-7f27-4520-b985-70095693efe6\") " pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.545720 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-668ss\" (UniqueName: \"kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss\") pod \"openstack-operator-index-dzdn2\" (UID: \"9dda7a38-7f27-4520-b985-70095693efe6\") " pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.571962 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-668ss\" (UniqueName: \"kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss\") pod \"openstack-operator-index-dzdn2\" (UID: \"9dda7a38-7f27-4520-b985-70095693efe6\") " pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:49 crc kubenswrapper[4888]: I1124 00:40:49.613954 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:50 crc kubenswrapper[4888]: I1124 00:40:50.078205 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:50 crc kubenswrapper[4888]: W1124 00:40:50.085727 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dda7a38_7f27_4520_b985_70095693efe6.slice/crio-b6412b39483b0c0907781aa33c03541d8cb6a947c74366e962ab706d4af2a5ce WatchSource:0}: Error finding container b6412b39483b0c0907781aa33c03541d8cb6a947c74366e962ab706d4af2a5ce: Status 404 returned error can't find the container with id b6412b39483b0c0907781aa33c03541d8cb6a947c74366e962ab706d4af2a5ce Nov 24 00:40:50 crc kubenswrapper[4888]: I1124 00:40:50.127561 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:50 crc kubenswrapper[4888]: I1124 00:40:50.191495 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:50 crc kubenswrapper[4888]: I1124 00:40:50.291277 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzdn2" event={"ID":"9dda7a38-7f27-4520-b985-70095693efe6","Type":"ContainerStarted","Data":"b6412b39483b0c0907781aa33c03541d8cb6a947c74366e962ab706d4af2a5ce"} Nov 24 00:40:51 crc kubenswrapper[4888]: I1124 00:40:51.580240 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.003748 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ggr8j"] Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.005087 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.014841 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ggr8j"] Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.101461 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q8g7\" (UniqueName: \"kubernetes.io/projected/927e1330-0a2c-4e33-b97d-5de99ce8138c-kube-api-access-9q8g7\") pod \"openstack-operator-index-ggr8j\" (UID: \"927e1330-0a2c-4e33-b97d-5de99ce8138c\") " pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.202970 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q8g7\" (UniqueName: \"kubernetes.io/projected/927e1330-0a2c-4e33-b97d-5de99ce8138c-kube-api-access-9q8g7\") pod \"openstack-operator-index-ggr8j\" (UID: \"927e1330-0a2c-4e33-b97d-5de99ce8138c\") " pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.222666 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q8g7\" (UniqueName: \"kubernetes.io/projected/927e1330-0a2c-4e33-b97d-5de99ce8138c-kube-api-access-9q8g7\") pod \"openstack-operator-index-ggr8j\" (UID: \"927e1330-0a2c-4e33-b97d-5de99ce8138c\") " pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.327604 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:40:52 crc kubenswrapper[4888]: I1124 00:40:52.844195 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ggr8j"] Nov 24 00:40:52 crc kubenswrapper[4888]: W1124 00:40:52.853323 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod927e1330_0a2c_4e33_b97d_5de99ce8138c.slice/crio-40b6bd26cf1c5ae63d7577ac011cd2a94c4976ebc3135b05336a79cebb9d2db1 WatchSource:0}: Error finding container 40b6bd26cf1c5ae63d7577ac011cd2a94c4976ebc3135b05336a79cebb9d2db1: Status 404 returned error can't find the container with id 40b6bd26cf1c5ae63d7577ac011cd2a94c4976ebc3135b05336a79cebb9d2db1 Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.335278 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ggr8j" event={"ID":"927e1330-0a2c-4e33-b97d-5de99ce8138c","Type":"ContainerStarted","Data":"747c26c7c6ee914902e2a804a42a099b794decae4b0e3c549f68ac889daac174"} Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.335786 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ggr8j" event={"ID":"927e1330-0a2c-4e33-b97d-5de99ce8138c","Type":"ContainerStarted","Data":"40b6bd26cf1c5ae63d7577ac011cd2a94c4976ebc3135b05336a79cebb9d2db1"} Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.340517 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzdn2" event={"ID":"9dda7a38-7f27-4520-b985-70095693efe6","Type":"ContainerStarted","Data":"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0"} Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.340673 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-dzdn2" podUID="9dda7a38-7f27-4520-b985-70095693efe6" containerName="registry-server" containerID="cri-o://3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0" gracePeriod=2 Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.362864 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ggr8j" podStartSLOduration=2.314569316 podStartE2EDuration="2.362831682s" podCreationTimestamp="2025-11-24 00:40:51 +0000 UTC" firstStartedPulling="2025-11-24 00:40:52.860667307 +0000 UTC m=+955.443351361" lastFinishedPulling="2025-11-24 00:40:52.908929693 +0000 UTC m=+955.491613727" observedRunningTime="2025-11-24 00:40:53.354318083 +0000 UTC m=+955.937002127" watchObservedRunningTime="2025-11-24 00:40:53.362831682 +0000 UTC m=+955.945515736" Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.384393 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dzdn2" podStartSLOduration=2.167227505 podStartE2EDuration="4.384362097s" podCreationTimestamp="2025-11-24 00:40:49 +0000 UTC" firstStartedPulling="2025-11-24 00:40:50.090560004 +0000 UTC m=+952.673244058" lastFinishedPulling="2025-11-24 00:40:52.307694576 +0000 UTC m=+954.890378650" observedRunningTime="2025-11-24 00:40:53.382220987 +0000 UTC m=+955.964905051" watchObservedRunningTime="2025-11-24 00:40:53.384362097 +0000 UTC m=+955.967046151" Nov 24 00:40:53 crc kubenswrapper[4888]: I1124 00:40:53.894622 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.036108 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-668ss\" (UniqueName: \"kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss\") pod \"9dda7a38-7f27-4520-b985-70095693efe6\" (UID: \"9dda7a38-7f27-4520-b985-70095693efe6\") " Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.043027 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss" (OuterVolumeSpecName: "kube-api-access-668ss") pod "9dda7a38-7f27-4520-b985-70095693efe6" (UID: "9dda7a38-7f27-4520-b985-70095693efe6"). InnerVolumeSpecName "kube-api-access-668ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.139851 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-668ss\" (UniqueName: \"kubernetes.io/projected/9dda7a38-7f27-4520-b985-70095693efe6-kube-api-access-668ss\") on node \"crc\" DevicePath \"\"" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.351617 4888 generic.go:334] "Generic (PLEG): container finished" podID="9dda7a38-7f27-4520-b985-70095693efe6" containerID="3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0" exitCode=0 Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.351760 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzdn2" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.351746 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzdn2" event={"ID":"9dda7a38-7f27-4520-b985-70095693efe6","Type":"ContainerDied","Data":"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0"} Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.352078 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzdn2" event={"ID":"9dda7a38-7f27-4520-b985-70095693efe6","Type":"ContainerDied","Data":"b6412b39483b0c0907781aa33c03541d8cb6a947c74366e962ab706d4af2a5ce"} Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.352130 4888 scope.go:117] "RemoveContainer" containerID="3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.390647 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.395493 4888 scope.go:117] "RemoveContainer" containerID="3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0" Nov 24 00:40:54 crc kubenswrapper[4888]: E1124 00:40:54.396828 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0\": container with ID starting with 3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0 not found: ID does not exist" containerID="3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.396898 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0"} err="failed to get container status \"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0\": rpc error: code = NotFound desc = could not find container \"3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0\": container with ID starting with 3c99e56b8ba7e95d523e8654bdb7572396ce436bd5a6aec0bed44b3d38ecd4f0 not found: ID does not exist" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.400686 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-dzdn2"] Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.625499 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zvfv2" Nov 24 00:40:54 crc kubenswrapper[4888]: I1124 00:40:54.701857 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-77rp4" Nov 24 00:40:55 crc kubenswrapper[4888]: I1124 00:40:55.132741 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bdrm6" Nov 24 00:40:56 crc kubenswrapper[4888]: I1124 00:40:56.264601 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dda7a38-7f27-4520-b985-70095693efe6" path="/var/lib/kubelet/pods/9dda7a38-7f27-4520-b985-70095693efe6/volumes" Nov 24 00:41:02 crc kubenswrapper[4888]: I1124 00:41:02.327985 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:41:02 crc kubenswrapper[4888]: I1124 00:41:02.328617 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:41:02 crc kubenswrapper[4888]: I1124 00:41:02.374527 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:41:02 crc kubenswrapper[4888]: I1124 00:41:02.536083 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-ggr8j" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.107661 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv"] Nov 24 00:41:22 crc kubenswrapper[4888]: E1124 00:41:22.108400 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dda7a38-7f27-4520-b985-70095693efe6" containerName="registry-server" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.108415 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dda7a38-7f27-4520-b985-70095693efe6" containerName="registry-server" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.108542 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dda7a38-7f27-4520-b985-70095693efe6" containerName="registry-server" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.109616 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.113293 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mhh6h" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.131900 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8v6l\" (UniqueName: \"kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.131992 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.132054 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.135153 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv"] Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.233654 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.234077 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8v6l\" (UniqueName: \"kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.234217 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.234391 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.234908 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.256316 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8v6l\" (UniqueName: \"kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l\") pod \"af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.426277 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:22 crc kubenswrapper[4888]: I1124 00:41:22.907660 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv"] Nov 24 00:41:23 crc kubenswrapper[4888]: I1124 00:41:23.702711 4888 generic.go:334] "Generic (PLEG): container finished" podID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerID="4029225b230434eba40311a9e61547abf7bc9b4a9e44198839542e5807b348fe" exitCode=0 Nov 24 00:41:23 crc kubenswrapper[4888]: I1124 00:41:23.702902 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" event={"ID":"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4","Type":"ContainerDied","Data":"4029225b230434eba40311a9e61547abf7bc9b4a9e44198839542e5807b348fe"} Nov 24 00:41:23 crc kubenswrapper[4888]: I1124 00:41:23.703227 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" event={"ID":"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4","Type":"ContainerStarted","Data":"7dc1393f1cbb89091356e2de1f5bd1085a85fb4309fde00ad43e5dde8b0d61b3"} Nov 24 00:41:24 crc kubenswrapper[4888]: I1124 00:41:24.711593 4888 generic.go:334] "Generic (PLEG): container finished" podID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerID="66126c64c05f893eba558c5d7fea9d5fe1206e29b18dfb66ee5a210da3a882f8" exitCode=0 Nov 24 00:41:24 crc kubenswrapper[4888]: I1124 00:41:24.711711 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" event={"ID":"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4","Type":"ContainerDied","Data":"66126c64c05f893eba558c5d7fea9d5fe1206e29b18dfb66ee5a210da3a882f8"} Nov 24 00:41:25 crc kubenswrapper[4888]: I1124 00:41:25.724013 4888 generic.go:334] "Generic (PLEG): container finished" podID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerID="9d7dde1dc4f9b7ab87ff349cf3808b67b32950dcf49d4b0efee098a3f7dfe33f" exitCode=0 Nov 24 00:41:25 crc kubenswrapper[4888]: I1124 00:41:25.724140 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" event={"ID":"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4","Type":"ContainerDied","Data":"9d7dde1dc4f9b7ab87ff349cf3808b67b32950dcf49d4b0efee098a3f7dfe33f"} Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.041009 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.229540 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle\") pod \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.229717 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util\") pod \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.229770 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8v6l\" (UniqueName: \"kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l\") pod \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\" (UID: \"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4\") " Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.230568 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle" (OuterVolumeSpecName: "bundle") pod "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" (UID: "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.235451 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l" (OuterVolumeSpecName: "kube-api-access-c8v6l") pod "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" (UID: "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4"). InnerVolumeSpecName "kube-api-access-c8v6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.243268 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util" (OuterVolumeSpecName: "util") pod "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" (UID: "de3cc12f-3ea8-4ea0-bbba-694b5a3258b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.332374 4888 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-util\") on node \"crc\" DevicePath \"\"" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.332433 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8v6l\" (UniqueName: \"kubernetes.io/projected/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-kube-api-access-c8v6l\") on node \"crc\" DevicePath \"\"" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.332456 4888 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de3cc12f-3ea8-4ea0-bbba-694b5a3258b4-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.746489 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" event={"ID":"de3cc12f-3ea8-4ea0-bbba-694b5a3258b4","Type":"ContainerDied","Data":"7dc1393f1cbb89091356e2de1f5bd1085a85fb4309fde00ad43e5dde8b0d61b3"} Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.746558 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dc1393f1cbb89091356e2de1f5bd1085a85fb4309fde00ad43e5dde8b0d61b3" Nov 24 00:41:27 crc kubenswrapper[4888]: I1124 00:41:27.746669 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.190224 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf"] Nov 24 00:41:34 crc kubenswrapper[4888]: E1124 00:41:34.191313 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="util" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.191327 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="util" Nov 24 00:41:34 crc kubenswrapper[4888]: E1124 00:41:34.191358 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="pull" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.191367 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="pull" Nov 24 00:41:34 crc kubenswrapper[4888]: E1124 00:41:34.191377 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="extract" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.191383 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="extract" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.191516 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3cc12f-3ea8-4ea0-bbba-694b5a3258b4" containerName="extract" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.192401 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.194779 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-d2rqg" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.230785 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf"] Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.372090 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx4ks\" (UniqueName: \"kubernetes.io/projected/9cc98d14-8e44-4556-b00a-ed2069d395cf-kube-api-access-wx4ks\") pod \"openstack-operator-controller-operator-75b7b5b9df-7j8cf\" (UID: \"9cc98d14-8e44-4556-b00a-ed2069d395cf\") " pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.474219 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx4ks\" (UniqueName: \"kubernetes.io/projected/9cc98d14-8e44-4556-b00a-ed2069d395cf-kube-api-access-wx4ks\") pod \"openstack-operator-controller-operator-75b7b5b9df-7j8cf\" (UID: \"9cc98d14-8e44-4556-b00a-ed2069d395cf\") " pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.539050 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx4ks\" (UniqueName: \"kubernetes.io/projected/9cc98d14-8e44-4556-b00a-ed2069d395cf-kube-api-access-wx4ks\") pod \"openstack-operator-controller-operator-75b7b5b9df-7j8cf\" (UID: \"9cc98d14-8e44-4556-b00a-ed2069d395cf\") " pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:34 crc kubenswrapper[4888]: I1124 00:41:34.811424 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:35 crc kubenswrapper[4888]: I1124 00:41:35.335140 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf"] Nov 24 00:41:35 crc kubenswrapper[4888]: I1124 00:41:35.827340 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" event={"ID":"9cc98d14-8e44-4556-b00a-ed2069d395cf","Type":"ContainerStarted","Data":"b7506f1ac1331b4ae55415cf6fd6baf12d97fd11464f86d9bb632f052b8249ad"} Nov 24 00:41:40 crc kubenswrapper[4888]: I1124 00:41:40.890620 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" event={"ID":"9cc98d14-8e44-4556-b00a-ed2069d395cf","Type":"ContainerStarted","Data":"ce6a1bd69c247f41d3548210dc9c76f38054badd70013a75eeeda337b7ec0936"} Nov 24 00:41:42 crc kubenswrapper[4888]: I1124 00:41:42.911266 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" event={"ID":"9cc98d14-8e44-4556-b00a-ed2069d395cf","Type":"ContainerStarted","Data":"e18750e919624015abbf1d330fa98c80794640e233966a88d1a7ca01e829bd8e"} Nov 24 00:41:42 crc kubenswrapper[4888]: I1124 00:41:42.911807 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:41:42 crc kubenswrapper[4888]: I1124 00:41:42.963931 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" podStartSLOduration=2.250263992 podStartE2EDuration="8.963898008s" podCreationTimestamp="2025-11-24 00:41:34 +0000 UTC" firstStartedPulling="2025-11-24 00:41:35.335967132 +0000 UTC m=+997.918651176" lastFinishedPulling="2025-11-24 00:41:42.049601148 +0000 UTC m=+1004.632285192" observedRunningTime="2025-11-24 00:41:42.95073061 +0000 UTC m=+1005.533414684" watchObservedRunningTime="2025-11-24 00:41:42.963898008 +0000 UTC m=+1005.546582082" Nov 24 00:41:44 crc kubenswrapper[4888]: I1124 00:41:44.816751 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-75b7b5b9df-7j8cf" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.210889 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.212848 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.214835 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2h7hf" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.224906 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.235694 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.235878 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.238820 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.239063 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-gx44p" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.242001 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vhr5c" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.288522 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.325921 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znk92\" (UniqueName: \"kubernetes.io/projected/5fe80165-08e7-4a34-8290-fd0d7fd22dd2-kube-api-access-znk92\") pod \"cinder-operator-controller-manager-6498cbf48f-4gh28\" (UID: \"5fe80165-08e7-4a34-8290-fd0d7fd22dd2\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.326288 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfbjb\" (UniqueName: \"kubernetes.io/projected/e7342ffc-1f02-4eca-9be9-5816f5f0c86c-kube-api-access-lfbjb\") pod \"barbican-operator-controller-manager-75fb479bcc-gphtr\" (UID: \"e7342ffc-1f02-4eca-9be9-5816f5f0c86c\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.326376 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpp4n\" (UniqueName: \"kubernetes.io/projected/a277e2cc-4990-4d79-8a4f-ec03c81d747d-kube-api-access-cpp4n\") pod \"designate-operator-controller-manager-767ccfd65f-9nb4l\" (UID: \"a277e2cc-4990-4d79-8a4f-ec03c81d747d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.334358 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.336107 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.340887 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.342383 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.344563 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ml467" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.350457 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dldf2" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.357106 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.369943 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.397900 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.407793 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.409121 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.419755 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.421206 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.424996 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.425100 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-zdt2q" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.425421 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mkfsg" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.430888 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.432742 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.432875 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22nzm\" (UniqueName: \"kubernetes.io/projected/6cf992a1-eac0-4269-b8bb-e14554116727-kube-api-access-22nzm\") pod \"heat-operator-controller-manager-56f54d6746-wbxrh\" (UID: \"6cf992a1-eac0-4269-b8bb-e14554116727\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.432961 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c6rz\" (UniqueName: \"kubernetes.io/projected/a701224c-a2ad-49b6-98b7-55fd37051e5b-kube-api-access-9c6rz\") pod \"glance-operator-controller-manager-7969689c84-mtqr5\" (UID: \"a701224c-a2ad-49b6-98b7-55fd37051e5b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.433013 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znk92\" (UniqueName: \"kubernetes.io/projected/5fe80165-08e7-4a34-8290-fd0d7fd22dd2-kube-api-access-znk92\") pod \"cinder-operator-controller-manager-6498cbf48f-4gh28\" (UID: \"5fe80165-08e7-4a34-8290-fd0d7fd22dd2\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.433044 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfbjb\" (UniqueName: \"kubernetes.io/projected/e7342ffc-1f02-4eca-9be9-5816f5f0c86c-kube-api-access-lfbjb\") pod \"barbican-operator-controller-manager-75fb479bcc-gphtr\" (UID: \"e7342ffc-1f02-4eca-9be9-5816f5f0c86c\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.433065 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpp4n\" (UniqueName: \"kubernetes.io/projected/a277e2cc-4990-4d79-8a4f-ec03c81d747d-kube-api-access-cpp4n\") pod \"designate-operator-controller-manager-767ccfd65f-9nb4l\" (UID: \"a277e2cc-4990-4d79-8a4f-ec03c81d747d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.445408 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5xg6h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.447754 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.457627 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.468008 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.479684 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfbjb\" (UniqueName: \"kubernetes.io/projected/e7342ffc-1f02-4eca-9be9-5816f5f0c86c-kube-api-access-lfbjb\") pod \"barbican-operator-controller-manager-75fb479bcc-gphtr\" (UID: \"e7342ffc-1f02-4eca-9be9-5816f5f0c86c\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.485212 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znk92\" (UniqueName: \"kubernetes.io/projected/5fe80165-08e7-4a34-8290-fd0d7fd22dd2-kube-api-access-znk92\") pod \"cinder-operator-controller-manager-6498cbf48f-4gh28\" (UID: \"5fe80165-08e7-4a34-8290-fd0d7fd22dd2\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.488397 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpp4n\" (UniqueName: \"kubernetes.io/projected/a277e2cc-4990-4d79-8a4f-ec03c81d747d-kube-api-access-cpp4n\") pod \"designate-operator-controller-manager-767ccfd65f-9nb4l\" (UID: \"a277e2cc-4990-4d79-8a4f-ec03c81d747d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.488487 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.538564 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.539756 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.541871 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542511 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7sd8\" (UniqueName: \"kubernetes.io/projected/3766dfec-2de1-4df2-8609-0cbbe8599342-kube-api-access-s7sd8\") pod \"ironic-operator-controller-manager-99b499f4-jk6ph\" (UID: \"3766dfec-2de1-4df2-8609-0cbbe8599342\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542564 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22nzm\" (UniqueName: \"kubernetes.io/projected/6cf992a1-eac0-4269-b8bb-e14554116727-kube-api-access-22nzm\") pod \"heat-operator-controller-manager-56f54d6746-wbxrh\" (UID: \"6cf992a1-eac0-4269-b8bb-e14554116727\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542607 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfbhf\" (UniqueName: \"kubernetes.io/projected/10a23051-0913-44fd-abda-89436a7e3b59-kube-api-access-nfbhf\") pod \"horizon-operator-controller-manager-598f69df5d-llt2h\" (UID: \"10a23051-0913-44fd-abda-89436a7e3b59\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542648 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nspx\" (UniqueName: \"kubernetes.io/projected/e6420e55-4951-4546-b252-44c25d49a0c3-kube-api-access-5nspx\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542667 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c6rz\" (UniqueName: \"kubernetes.io/projected/a701224c-a2ad-49b6-98b7-55fd37051e5b-kube-api-access-9c6rz\") pod \"glance-operator-controller-manager-7969689c84-mtqr5\" (UID: \"a701224c-a2ad-49b6-98b7-55fd37051e5b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.542696 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.546166 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-zq4dw" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.561490 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.562268 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.563193 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.567430 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gfrmw" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.580681 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.582449 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22nzm\" (UniqueName: \"kubernetes.io/projected/6cf992a1-eac0-4269-b8bb-e14554116727-kube-api-access-22nzm\") pod \"heat-operator-controller-manager-56f54d6746-wbxrh\" (UID: \"6cf992a1-eac0-4269-b8bb-e14554116727\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.584328 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.597858 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c6rz\" (UniqueName: \"kubernetes.io/projected/a701224c-a2ad-49b6-98b7-55fd37051e5b-kube-api-access-9c6rz\") pod \"glance-operator-controller-manager-7969689c84-mtqr5\" (UID: \"a701224c-a2ad-49b6-98b7-55fd37051e5b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.601123 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.608136 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.609833 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.614157 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-sgcg5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.614380 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.616318 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.621797 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2zd96" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.635340 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.637231 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.649781 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfbhf\" (UniqueName: \"kubernetes.io/projected/10a23051-0913-44fd-abda-89436a7e3b59-kube-api-access-nfbhf\") pod \"horizon-operator-controller-manager-598f69df5d-llt2h\" (UID: \"10a23051-0913-44fd-abda-89436a7e3b59\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.649865 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nspx\" (UniqueName: \"kubernetes.io/projected/e6420e55-4951-4546-b252-44c25d49a0c3-kube-api-access-5nspx\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.649904 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.649955 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8xq\" (UniqueName: \"kubernetes.io/projected/587fee04-3081-4c59-8125-0e47436e2347-kube-api-access-jc8xq\") pod \"keystone-operator-controller-manager-7454b96578-8kqtd\" (UID: \"587fee04-3081-4c59-8125-0e47436e2347\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.649984 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7sd8\" (UniqueName: \"kubernetes.io/projected/3766dfec-2de1-4df2-8609-0cbbe8599342-kube-api-access-s7sd8\") pod \"ironic-operator-controller-manager-99b499f4-jk6ph\" (UID: \"3766dfec-2de1-4df2-8609-0cbbe8599342\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.650010 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz8h7\" (UniqueName: \"kubernetes.io/projected/c18797ad-a3dc-46af-9aaf-d9e17918228a-kube-api-access-hz8h7\") pod \"manila-operator-controller-manager-58f887965d-tsb9j\" (UID: \"c18797ad-a3dc-46af-9aaf-d9e17918228a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:09 crc kubenswrapper[4888]: E1124 00:42:09.650579 4888 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 00:42:09 crc kubenswrapper[4888]: E1124 00:42:09.650782 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert podName:e6420e55-4951-4546-b252-44c25d49a0c3 nodeName:}" failed. No retries permitted until 2025-11-24 00:42:10.150746433 +0000 UTC m=+1032.733430477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert") pod "infra-operator-controller-manager-6dd8864d7c-zl954" (UID: "e6420e55-4951-4546-b252-44c25d49a0c3") : secret "infra-operator-webhook-server-cert" not found Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.651129 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.663643 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-gldxm" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.680167 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.682554 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.695585 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfbhf\" (UniqueName: \"kubernetes.io/projected/10a23051-0913-44fd-abda-89436a7e3b59-kube-api-access-nfbhf\") pod \"horizon-operator-controller-manager-598f69df5d-llt2h\" (UID: \"10a23051-0913-44fd-abda-89436a7e3b59\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.696560 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nspx\" (UniqueName: \"kubernetes.io/projected/e6420e55-4951-4546-b252-44c25d49a0c3-kube-api-access-5nspx\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.701832 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.707264 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.726455 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.749823 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7sd8\" (UniqueName: \"kubernetes.io/projected/3766dfec-2de1-4df2-8609-0cbbe8599342-kube-api-access-s7sd8\") pod \"ironic-operator-controller-manager-99b499f4-jk6ph\" (UID: \"3766dfec-2de1-4df2-8609-0cbbe8599342\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.752013 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxp87\" (UniqueName: \"kubernetes.io/projected/f4bc41bd-b9d5-4acf-978e-75144d8b43a3-kube-api-access-cxp87\") pod \"nova-operator-controller-manager-cfbb9c588-7zjtv\" (UID: \"f4bc41bd-b9d5-4acf-978e-75144d8b43a3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.752089 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxjf4\" (UniqueName: \"kubernetes.io/projected/dbaba630-daf7-46c3-807a-6e76128fc7d6-kube-api-access-dxjf4\") pod \"mariadb-operator-controller-manager-54b5986bb8-4pmdz\" (UID: \"dbaba630-daf7-46c3-807a-6e76128fc7d6\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.752149 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgzbw\" (UniqueName: \"kubernetes.io/projected/afa14c5d-d820-4b05-bddd-4d00e49a421a-kube-api-access-fgzbw\") pod \"neutron-operator-controller-manager-78bd47f458-fn4cs\" (UID: \"afa14c5d-d820-4b05-bddd-4d00e49a421a\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.752186 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc8xq\" (UniqueName: \"kubernetes.io/projected/587fee04-3081-4c59-8125-0e47436e2347-kube-api-access-jc8xq\") pod \"keystone-operator-controller-manager-7454b96578-8kqtd\" (UID: \"587fee04-3081-4c59-8125-0e47436e2347\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.752221 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz8h7\" (UniqueName: \"kubernetes.io/projected/c18797ad-a3dc-46af-9aaf-d9e17918228a-kube-api-access-hz8h7\") pod \"manila-operator-controller-manager-58f887965d-tsb9j\" (UID: \"c18797ad-a3dc-46af-9aaf-d9e17918228a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.757905 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.758000 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.762088 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.763671 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.769746 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rwcm9" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.772485 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5h6qb" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.772669 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.774573 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz8h7\" (UniqueName: \"kubernetes.io/projected/c18797ad-a3dc-46af-9aaf-d9e17918228a-kube-api-access-hz8h7\") pod \"manila-operator-controller-manager-58f887965d-tsb9j\" (UID: \"c18797ad-a3dc-46af-9aaf-d9e17918228a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.785099 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc8xq\" (UniqueName: \"kubernetes.io/projected/587fee04-3081-4c59-8125-0e47436e2347-kube-api-access-jc8xq\") pod \"keystone-operator-controller-manager-7454b96578-8kqtd\" (UID: \"587fee04-3081-4c59-8125-0e47436e2347\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.846492 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.862746 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxp87\" (UniqueName: \"kubernetes.io/projected/f4bc41bd-b9d5-4acf-978e-75144d8b43a3-kube-api-access-cxp87\") pod \"nova-operator-controller-manager-cfbb9c588-7zjtv\" (UID: \"f4bc41bd-b9d5-4acf-978e-75144d8b43a3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.863588 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mjhc\" (UniqueName: \"kubernetes.io/projected/750c449f-a775-4ee8-82b8-a6be531250fe-kube-api-access-9mjhc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-vxffn\" (UID: \"750c449f-a775-4ee8-82b8-a6be531250fe\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.863741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4t9h\" (UniqueName: \"kubernetes.io/projected/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-kube-api-access-v4t9h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.863956 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxjf4\" (UniqueName: \"kubernetes.io/projected/dbaba630-daf7-46c3-807a-6e76128fc7d6-kube-api-access-dxjf4\") pod \"mariadb-operator-controller-manager-54b5986bb8-4pmdz\" (UID: \"dbaba630-daf7-46c3-807a-6e76128fc7d6\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.864233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgzbw\" (UniqueName: \"kubernetes.io/projected/afa14c5d-d820-4b05-bddd-4d00e49a421a-kube-api-access-fgzbw\") pod \"neutron-operator-controller-manager-78bd47f458-fn4cs\" (UID: \"afa14c5d-d820-4b05-bddd-4d00e49a421a\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.864350 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.867884 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.872570 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.897179 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.910994 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxjf4\" (UniqueName: \"kubernetes.io/projected/dbaba630-daf7-46c3-807a-6e76128fc7d6-kube-api-access-dxjf4\") pod \"mariadb-operator-controller-manager-54b5986bb8-4pmdz\" (UID: \"dbaba630-daf7-46c3-807a-6e76128fc7d6\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.926594 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgzbw\" (UniqueName: \"kubernetes.io/projected/afa14c5d-d820-4b05-bddd-4d00e49a421a-kube-api-access-fgzbw\") pod \"neutron-operator-controller-manager-78bd47f458-fn4cs\" (UID: \"afa14c5d-d820-4b05-bddd-4d00e49a421a\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.926910 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-j9v2l" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.960651 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh"] Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.968103 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxp87\" (UniqueName: \"kubernetes.io/projected/f4bc41bd-b9d5-4acf-978e-75144d8b43a3-kube-api-access-cxp87\") pod \"nova-operator-controller-manager-cfbb9c588-7zjtv\" (UID: \"f4bc41bd-b9d5-4acf-978e-75144d8b43a3\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.968566 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:09 crc kubenswrapper[4888]: E1124 00:42:09.968720 4888 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.968799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-587bs\" (UniqueName: \"kubernetes.io/projected/8a57fb36-c0db-4778-9c17-7241b5e8f602-kube-api-access-587bs\") pod \"ovn-operator-controller-manager-54fc5f65b7-c9cfn\" (UID: \"8a57fb36-c0db-4778-9c17-7241b5e8f602\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:09 crc kubenswrapper[4888]: E1124 00:42:09.968975 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert podName:2b35fb96-9ea1-4aaa-9396-cd5f705f63bb nodeName:}" failed. No retries permitted until 2025-11-24 00:42:10.468958062 +0000 UTC m=+1033.051642106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" (UID: "2b35fb96-9ea1-4aaa-9396-cd5f705f63bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.969322 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mjhc\" (UniqueName: \"kubernetes.io/projected/750c449f-a775-4ee8-82b8-a6be531250fe-kube-api-access-9mjhc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-vxffn\" (UID: \"750c449f-a775-4ee8-82b8-a6be531250fe\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.969400 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4t9h\" (UniqueName: \"kubernetes.io/projected/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-kube-api-access-v4t9h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:09 crc kubenswrapper[4888]: I1124 00:42:09.984981 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.022533 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.031788 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.037370 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4t9h\" (UniqueName: \"kubernetes.io/projected/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-kube-api-access-v4t9h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.039143 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mjhc\" (UniqueName: \"kubernetes.io/projected/750c449f-a775-4ee8-82b8-a6be531250fe-kube-api-access-9mjhc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-vxffn\" (UID: \"750c449f-a775-4ee8-82b8-a6be531250fe\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.041554 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.053448 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.054106 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.055166 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.059321 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7rjzq" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.068916 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.077717 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.080614 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkh9\" (UniqueName: \"kubernetes.io/projected/fef147fd-ee85-4c05-9652-5e44223a3e5e-kube-api-access-wtkh9\") pod \"placement-operator-controller-manager-5b797b8dff-98b6n\" (UID: \"fef147fd-ee85-4c05-9652-5e44223a3e5e\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.080736 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.080786 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-587bs\" (UniqueName: \"kubernetes.io/projected/8a57fb36-c0db-4778-9c17-7241b5e8f602-kube-api-access-587bs\") pod \"ovn-operator-controller-manager-54fc5f65b7-c9cfn\" (UID: \"8a57fb36-c0db-4778-9c17-7241b5e8f602\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.084953 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-m77qm" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.119546 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.121582 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.126095 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-vgwwd" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.132936 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.137018 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-587bs\" (UniqueName: \"kubernetes.io/projected/8a57fb36-c0db-4778-9c17-7241b5e8f602-kube-api-access-587bs\") pod \"ovn-operator-controller-manager-54fc5f65b7-c9cfn\" (UID: \"8a57fb36-c0db-4778-9c17-7241b5e8f602\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.141348 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.166458 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.180435 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-mn678"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.182267 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.182759 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dp7z\" (UniqueName: \"kubernetes.io/projected/3a5b66a7-1cf6-4840-870c-f7a6968bd08a-kube-api-access-9dp7z\") pod \"swift-operator-controller-manager-d656998f4-5ck2h\" (UID: \"3a5b66a7-1cf6-4840-870c-f7a6968bd08a\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.182890 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkh9\" (UniqueName: \"kubernetes.io/projected/fef147fd-ee85-4c05-9652-5e44223a3e5e-kube-api-access-wtkh9\") pod \"placement-operator-controller-manager-5b797b8dff-98b6n\" (UID: \"fef147fd-ee85-4c05-9652-5e44223a3e5e\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.182923 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.182968 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n84v7\" (UniqueName: \"kubernetes.io/projected/bc56a88d-e719-40f5-a210-1b7afdffae8b-kube-api-access-n84v7\") pod \"telemetry-operator-controller-manager-67c4db9997-7rqlj\" (UID: \"bc56a88d-e719-40f5-a210-1b7afdffae8b\") " pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.189729 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4bq4n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.200735 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6420e55-4951-4546-b252-44c25d49a0c3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-zl954\" (UID: \"e6420e55-4951-4546-b252-44c25d49a0c3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.215912 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.222989 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-mn678"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.236629 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkh9\" (UniqueName: \"kubernetes.io/projected/fef147fd-ee85-4c05-9652-5e44223a3e5e-kube-api-access-wtkh9\") pod \"placement-operator-controller-manager-5b797b8dff-98b6n\" (UID: \"fef147fd-ee85-4c05-9652-5e44223a3e5e\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.254300 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.255607 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.256971 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.260204 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.270046 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-drs59" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.297509 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dp7z\" (UniqueName: \"kubernetes.io/projected/3a5b66a7-1cf6-4840-870c-f7a6968bd08a-kube-api-access-9dp7z\") pod \"swift-operator-controller-manager-d656998f4-5ck2h\" (UID: \"3a5b66a7-1cf6-4840-870c-f7a6968bd08a\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.298896 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dc6z\" (UniqueName: \"kubernetes.io/projected/c36e3742-c48a-48d1-96c8-87c3809a4183-kube-api-access-6dc6z\") pod \"watcher-operator-controller-manager-8c6448b9f-6mmtv\" (UID: \"c36e3742-c48a-48d1-96c8-87c3809a4183\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.298981 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjfn8\" (UniqueName: \"kubernetes.io/projected/82f2609b-e676-49f8-9ab4-d06fb3e626be-kube-api-access-hjfn8\") pod \"test-operator-controller-manager-b4c496f69-mn678\" (UID: \"82f2609b-e676-49f8-9ab4-d06fb3e626be\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.299005 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n84v7\" (UniqueName: \"kubernetes.io/projected/bc56a88d-e719-40f5-a210-1b7afdffae8b-kube-api-access-n84v7\") pod \"telemetry-operator-controller-manager-67c4db9997-7rqlj\" (UID: \"bc56a88d-e719-40f5-a210-1b7afdffae8b\") " pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.350311 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n84v7\" (UniqueName: \"kubernetes.io/projected/bc56a88d-e719-40f5-a210-1b7afdffae8b-kube-api-access-n84v7\") pod \"telemetry-operator-controller-manager-67c4db9997-7rqlj\" (UID: \"bc56a88d-e719-40f5-a210-1b7afdffae8b\") " pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.369672 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.416296 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dc6z\" (UniqueName: \"kubernetes.io/projected/c36e3742-c48a-48d1-96c8-87c3809a4183-kube-api-access-6dc6z\") pod \"watcher-operator-controller-manager-8c6448b9f-6mmtv\" (UID: \"c36e3742-c48a-48d1-96c8-87c3809a4183\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.416575 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjfn8\" (UniqueName: \"kubernetes.io/projected/82f2609b-e676-49f8-9ab4-d06fb3e626be-kube-api-access-hjfn8\") pod \"test-operator-controller-manager-b4c496f69-mn678\" (UID: \"82f2609b-e676-49f8-9ab4-d06fb3e626be\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.446890 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dp7z\" (UniqueName: \"kubernetes.io/projected/3a5b66a7-1cf6-4840-870c-f7a6968bd08a-kube-api-access-9dp7z\") pod \"swift-operator-controller-manager-d656998f4-5ck2h\" (UID: \"3a5b66a7-1cf6-4840-870c-f7a6968bd08a\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.456616 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dc6z\" (UniqueName: \"kubernetes.io/projected/c36e3742-c48a-48d1-96c8-87c3809a4183-kube-api-access-6dc6z\") pod \"watcher-operator-controller-manager-8c6448b9f-6mmtv\" (UID: \"c36e3742-c48a-48d1-96c8-87c3809a4183\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.457122 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjfn8\" (UniqueName: \"kubernetes.io/projected/82f2609b-e676-49f8-9ab4-d06fb3e626be-kube-api-access-hjfn8\") pod \"test-operator-controller-manager-b4c496f69-mn678\" (UID: \"82f2609b-e676-49f8-9ab4-d06fb3e626be\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.476636 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.476706 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.478272 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.506346 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.506924 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-h5g5k" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.518143 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2mx7\" (UniqueName: \"kubernetes.io/projected/db75890a-4c13-4f98-a410-aa4f98890ccc-kube-api-access-d2mx7\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.518224 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.518259 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:10 crc kubenswrapper[4888]: E1124 00:42:10.518418 4888 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 00:42:10 crc kubenswrapper[4888]: E1124 00:42:10.518484 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert podName:2b35fb96-9ea1-4aaa-9396-cd5f705f63bb nodeName:}" failed. No retries permitted until 2025-11-24 00:42:11.51846488 +0000 UTC m=+1034.101148924 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" (UID: "2b35fb96-9ea1-4aaa-9396-cd5f705f63bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.543890 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.579663 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.604881 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.607014 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.616163 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9wb8g" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.618923 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.620870 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9db4\" (UniqueName: \"kubernetes.io/projected/e569ce1f-9272-4c6a-bb91-a01616bb5883-kube-api-access-s9db4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv\" (UID: \"e569ce1f-9272-4c6a-bb91-a01616bb5883\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.621010 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2mx7\" (UniqueName: \"kubernetes.io/projected/db75890a-4c13-4f98-a410-aa4f98890ccc-kube-api-access-d2mx7\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.621118 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: E1124 00:42:10.621507 4888 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 00:42:10 crc kubenswrapper[4888]: E1124 00:42:10.621654 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert podName:db75890a-4c13-4f98-a410-aa4f98890ccc nodeName:}" failed. No retries permitted until 2025-11-24 00:42:11.121631485 +0000 UTC m=+1033.704315529 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert") pod "openstack-operator-controller-manager-6d84888d56-jk5d2" (UID: "db75890a-4c13-4f98-a410-aa4f98890ccc") : secret "webhook-server-cert" not found Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.637574 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.673442 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.678647 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2mx7\" (UniqueName: \"kubernetes.io/projected/db75890a-4c13-4f98-a410-aa4f98890ccc-kube-api-access-d2mx7\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.723644 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9db4\" (UniqueName: \"kubernetes.io/projected/e569ce1f-9272-4c6a-bb91-a01616bb5883-kube-api-access-s9db4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv\" (UID: \"e569ce1f-9272-4c6a-bb91-a01616bb5883\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.725030 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.738462 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.746942 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr"] Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.748716 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9db4\" (UniqueName: \"kubernetes.io/projected/e569ce1f-9272-4c6a-bb91-a01616bb5883-kube-api-access-s9db4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv\" (UID: \"e569ce1f-9272-4c6a-bb91-a01616bb5883\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.750132 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.827237 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" Nov 24 00:42:10 crc kubenswrapper[4888]: I1124 00:42:10.907395 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5"] Nov 24 00:42:10 crc kubenswrapper[4888]: W1124 00:42:10.953106 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda701224c_a2ad_49b6_98b7_55fd37051e5b.slice/crio-3324a190125443cbf30c742ca3fb0be5f71b265908d9f4a15db38a6494fe993f WatchSource:0}: Error finding container 3324a190125443cbf30c742ca3fb0be5f71b265908d9f4a15db38a6494fe993f: Status 404 returned error can't find the container with id 3324a190125443cbf30c742ca3fb0be5f71b265908d9f4a15db38a6494fe993f Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.115028 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.124311 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.131529 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.138383 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.145718 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db75890a-4c13-4f98-a410-aa4f98890ccc-cert\") pod \"openstack-operator-controller-manager-6d84888d56-jk5d2\" (UID: \"db75890a-4c13-4f98-a410-aa4f98890ccc\") " pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.168029 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" event={"ID":"a701224c-a2ad-49b6-98b7-55fd37051e5b","Type":"ContainerStarted","Data":"3324a190125443cbf30c742ca3fb0be5f71b265908d9f4a15db38a6494fe993f"} Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.169426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" event={"ID":"a277e2cc-4990-4d79-8a4f-ec03c81d747d","Type":"ContainerStarted","Data":"2f35e8c45975968dc02260b5c35180e790957436e15f642aca5ff67671c2555b"} Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.174860 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" event={"ID":"e7342ffc-1f02-4eca-9be9-5816f5f0c86c","Type":"ContainerStarted","Data":"d1d86dcc7b0a15f6e47c917d71e24e1d411a57bad6e89aa0bfa15c12fc1a9fdf"} Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.177798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" event={"ID":"5fe80165-08e7-4a34-8290-fd0d7fd22dd2","Type":"ContainerStarted","Data":"8ccda2e3fe7bd652d912ab21bfad132787e6c9f4aa8dd97fac51dc993166f5dc"} Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.396575 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.413363 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.433673 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j"] Nov 24 00:42:11 crc kubenswrapper[4888]: W1124 00:42:11.436413 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafa14c5d_d820_4b05_bddd_4d00e49a421a.slice/crio-e89845d5302ef9c34801b4451b328bf45661fb171dec07e3356c2b7af8470d47 WatchSource:0}: Error finding container e89845d5302ef9c34801b4451b328bf45661fb171dec07e3356c2b7af8470d47: Status 404 returned error can't find the container with id e89845d5302ef9c34801b4451b328bf45661fb171dec07e3356c2b7af8470d47 Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.440258 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv"] Nov 24 00:42:11 crc kubenswrapper[4888]: W1124 00:42:11.445087 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4bc41bd_b9d5_4acf_978e_75144d8b43a3.slice/crio-906b4ea623e1d8fdb6dc1bd46fe0168903337a0b501aa43dc57a7ebe87dd95c7 WatchSource:0}: Error finding container 906b4ea623e1d8fdb6dc1bd46fe0168903337a0b501aa43dc57a7ebe87dd95c7: Status 404 returned error can't find the container with id 906b4ea623e1d8fdb6dc1bd46fe0168903337a0b501aa43dc57a7ebe87dd95c7 Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.545564 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.551048 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b35fb96-9ea1-4aaa-9396-cd5f705f63bb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh\" (UID: \"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.597161 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.702143 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.870568 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.885658 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.902378 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn"] Nov 24 00:42:11 crc kubenswrapper[4888]: W1124 00:42:11.912234 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a57fb36_c0db_4778_9c17_7241b5e8f602.slice/crio-34a3334afdceca9dfa6ec0e0bf9ca69dd9b5d45af24ba04305e62ad096b765bd WatchSource:0}: Error finding container 34a3334afdceca9dfa6ec0e0bf9ca69dd9b5d45af24ba04305e62ad096b765bd: Status 404 returned error can't find the container with id 34a3334afdceca9dfa6ec0e0bf9ca69dd9b5d45af24ba04305e62ad096b765bd Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.920802 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv"] Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.930778 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9db4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv_openstack-operators(e569ce1f-9272-4c6a-bb91-a01616bb5883): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.934030 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" podUID="e569ce1f-9272-4c6a-bb91-a01616bb5883" Nov 24 00:42:11 crc kubenswrapper[4888]: W1124 00:42:11.936395 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a5b66a7_1cf6_4840_870c_f7a6968bd08a.slice/crio-8f436e5106ef7020bfb941bbb01580a9b3e93e967c2b9eb7030d36330349a93a WatchSource:0}: Error finding container 8f436e5106ef7020bfb941bbb01580a9b3e93e967c2b9eb7030d36330349a93a: Status 404 returned error can't find the container with id 8f436e5106ef7020bfb941bbb01580a9b3e93e967c2b9eb7030d36330349a93a Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.936407 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954"] Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.936956 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mjhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-54cfbf4c7d-vxffn_openstack-operators(750c449f-a775-4ee8-82b8-a6be531250fe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.937738 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9dp7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-5ck2h_openstack-operators(3a5b66a7-1cf6-4840-870c-f7a6968bd08a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.943919 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.950002 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.954780 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj"] Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.962345 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.146:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n84v7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-67c4db9997-7rqlj_openstack-operators(bc56a88d-e719-40f5-a210-1b7afdffae8b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 00:42:11 crc kubenswrapper[4888]: E1124 00:42:11.962432 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjfn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-mn678_openstack-operators(82f2609b-e676-49f8-9ab4-d06fb3e626be): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.975752 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-mn678"] Nov 24 00:42:11 crc kubenswrapper[4888]: I1124 00:42:11.996885 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv"] Nov 24 00:42:12 crc kubenswrapper[4888]: W1124 00:42:12.051869 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc36e3742_c48a_48d1_96c8_87c3809a4183.slice/crio-a871c0596b24018c8fe38451cc0556457f38b827f1fd6037e00398eb007cb59d WatchSource:0}: Error finding container a871c0596b24018c8fe38451cc0556457f38b827f1fd6037e00398eb007cb59d: Status 404 returned error can't find the container with id a871c0596b24018c8fe38451cc0556457f38b827f1fd6037e00398eb007cb59d Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.059423 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2"] Nov 24 00:42:12 crc kubenswrapper[4888]: W1124 00:42:12.101611 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb75890a_4c13_4f98_a410_aa4f98890ccc.slice/crio-4b4e64fca53fd6b4b19c9530dd23593b88d1cdd442a9892893be12246cbeb9ab WatchSource:0}: Error finding container 4b4e64fca53fd6b4b19c9530dd23593b88d1cdd442a9892893be12246cbeb9ab: Status 404 returned error can't find the container with id 4b4e64fca53fd6b4b19c9530dd23593b88d1cdd442a9892893be12246cbeb9ab Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.215689 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" event={"ID":"e569ce1f-9272-4c6a-bb91-a01616bb5883","Type":"ContainerStarted","Data":"16dd831ed1f380ba8390d682225e1061fa2c7b69868a77628143cd1757b97524"} Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.220921 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" podUID="e569ce1f-9272-4c6a-bb91-a01616bb5883" Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.239365 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" event={"ID":"f4bc41bd-b9d5-4acf-978e-75144d8b43a3","Type":"ContainerStarted","Data":"906b4ea623e1d8fdb6dc1bd46fe0168903337a0b501aa43dc57a7ebe87dd95c7"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.286424 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" event={"ID":"bc56a88d-e719-40f5-a210-1b7afdffae8b","Type":"ContainerStarted","Data":"0e469e9e94c9de4fdab52390dd1bf13d2e298b3b8155749daae789b9a516aca4"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" event={"ID":"8a57fb36-c0db-4778-9c17-7241b5e8f602","Type":"ContainerStarted","Data":"34a3334afdceca9dfa6ec0e0bf9ca69dd9b5d45af24ba04305e62ad096b765bd"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" event={"ID":"db75890a-4c13-4f98-a410-aa4f98890ccc","Type":"ContainerStarted","Data":"4b4e64fca53fd6b4b19c9530dd23593b88d1cdd442a9892893be12246cbeb9ab"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287374 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" event={"ID":"587fee04-3081-4c59-8125-0e47436e2347","Type":"ContainerStarted","Data":"1d68136bac4d67577cf3ec25c887d86e291c1ab7413a392b60838af02e65bff6"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287455 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" event={"ID":"c36e3742-c48a-48d1-96c8-87c3809a4183","Type":"ContainerStarted","Data":"a871c0596b24018c8fe38451cc0556457f38b827f1fd6037e00398eb007cb59d"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287531 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" event={"ID":"fef147fd-ee85-4c05-9652-5e44223a3e5e","Type":"ContainerStarted","Data":"d1935b741736f4976a3601639df01657e74ee495b96d786d4ff884d110fab4c9"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287611 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" event={"ID":"3a5b66a7-1cf6-4840-870c-f7a6968bd08a","Type":"ContainerStarted","Data":"8f436e5106ef7020bfb941bbb01580a9b3e93e967c2b9eb7030d36330349a93a"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287688 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" event={"ID":"82f2609b-e676-49f8-9ab4-d06fb3e626be","Type":"ContainerStarted","Data":"a095090d9d4f1b7e63fa6a3e5410a3972c5b4ad06d66f5ad9d7d07ef0fc90697"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287763 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" event={"ID":"10a23051-0913-44fd-abda-89436a7e3b59","Type":"ContainerStarted","Data":"8e70d9567eefbe8eacb7fe70d82427a4b4e5d0e7415c80eb0f5d715d6e22b2f3"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.287855 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" event={"ID":"dbaba630-daf7-46c3-807a-6e76128fc7d6","Type":"ContainerStarted","Data":"52db0c4dda9a2141a4028223b862e72a7bf01ce5e730395b51da8abdc4fa1cf9"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.288746 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" event={"ID":"3766dfec-2de1-4df2-8609-0cbbe8599342","Type":"ContainerStarted","Data":"bb27672b86aa006f55eec24331df8368be7069bf5565297d3ebf934afc61d495"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.296495 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh"] Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.322561 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" event={"ID":"e6420e55-4951-4546-b252-44c25d49a0c3","Type":"ContainerStarted","Data":"422525cf06448611bbf2d7cc9dc9bf59f4c9e71f8514a3544a3b9b3ecad876a7"} Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.332569 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" podUID="750c449f-a775-4ee8-82b8-a6be531250fe" Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.335834 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" event={"ID":"afa14c5d-d820-4b05-bddd-4d00e49a421a","Type":"ContainerStarted","Data":"e89845d5302ef9c34801b4451b328bf45661fb171dec07e3356c2b7af8470d47"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.348282 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" event={"ID":"6cf992a1-eac0-4269-b8bb-e14554116727","Type":"ContainerStarted","Data":"32f9bc1154319788f86e844169d9c36aa9c602978ceca8b18fc205894a7c9e6d"} Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.358529 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" event={"ID":"750c449f-a775-4ee8-82b8-a6be531250fe","Type":"ContainerStarted","Data":"5a8e626d82ba397ccaef866948f4c26e079650ac0674dc148729a70538ef402b"} Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.361024 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" podUID="750c449f-a775-4ee8-82b8-a6be531250fe" Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.363129 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" podUID="82f2609b-e676-49f8-9ab4-d06fb3e626be" Nov 24 00:42:12 crc kubenswrapper[4888]: I1124 00:42:12.363238 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" event={"ID":"c18797ad-a3dc-46af-9aaf-d9e17918228a","Type":"ContainerStarted","Data":"af8589d7934b7d30e305017c87e74e3a29509a4715d287d57e07152e33268f5f"} Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.366210 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" podUID="3a5b66a7-1cf6-4840-870c-f7a6968bd08a" Nov 24 00:42:12 crc kubenswrapper[4888]: E1124 00:42:12.417361 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" podUID="bc56a88d-e719-40f5-a210-1b7afdffae8b" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.423329 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" event={"ID":"bc56a88d-e719-40f5-a210-1b7afdffae8b","Type":"ContainerStarted","Data":"d3822eba91072705759bc44bf12263507085cf5abdafc5384f2a3e68a879876a"} Nov 24 00:42:13 crc kubenswrapper[4888]: E1124 00:42:13.426892 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.146:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" podUID="bc56a88d-e719-40f5-a210-1b7afdffae8b" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.430184 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" event={"ID":"750c449f-a775-4ee8-82b8-a6be531250fe","Type":"ContainerStarted","Data":"7bbf12d961af7ec900b7c9ac79d45f506b50e3e1a344621a95c3d16f3cfcc53b"} Nov 24 00:42:13 crc kubenswrapper[4888]: E1124 00:42:13.435938 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" podUID="750c449f-a775-4ee8-82b8-a6be531250fe" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.445505 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" event={"ID":"db75890a-4c13-4f98-a410-aa4f98890ccc","Type":"ContainerStarted","Data":"e21aa32464affbcfd5bea85c7bf7494bc24a53bcddf05f4ff3f46eed8641aceb"} Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.445565 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" event={"ID":"db75890a-4c13-4f98-a410-aa4f98890ccc","Type":"ContainerStarted","Data":"495fce5505dab2bfbcefd0f833893286b15a98c2211150b602d2e6b3878c2961"} Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.446506 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.457377 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" event={"ID":"3a5b66a7-1cf6-4840-870c-f7a6968bd08a","Type":"ContainerStarted","Data":"7245508fa6548ecbc4fc0371c2bef62eee01fd9d186058866d62e89f9895e2c1"} Nov 24 00:42:13 crc kubenswrapper[4888]: E1124 00:42:13.467272 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" podUID="3a5b66a7-1cf6-4840-870c-f7a6968bd08a" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.485483 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" event={"ID":"82f2609b-e676-49f8-9ab4-d06fb3e626be","Type":"ContainerStarted","Data":"a92979b4fa658d7c0b1dbb0e076a0c4ec17bdc72403e929dcb82dc5caca7af5d"} Nov 24 00:42:13 crc kubenswrapper[4888]: E1124 00:42:13.492593 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" podUID="82f2609b-e676-49f8-9ab4-d06fb3e626be" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.500609 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" event={"ID":"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb","Type":"ContainerStarted","Data":"97ba5fac330aaa01509f44ff7397378d5e6def7b47ad39aeeec0d6badb063c03"} Nov 24 00:42:13 crc kubenswrapper[4888]: E1124 00:42:13.503271 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" podUID="e569ce1f-9272-4c6a-bb91-a01616bb5883" Nov 24 00:42:13 crc kubenswrapper[4888]: I1124 00:42:13.529479 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" podStartSLOduration=4.529454696 podStartE2EDuration="4.529454696s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:42:13.522617985 +0000 UTC m=+1036.105302029" watchObservedRunningTime="2025-11-24 00:42:13.529454696 +0000 UTC m=+1036.112138740" Nov 24 00:42:14 crc kubenswrapper[4888]: E1124 00:42:14.520318 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" podUID="750c449f-a775-4ee8-82b8-a6be531250fe" Nov 24 00:42:14 crc kubenswrapper[4888]: E1124 00:42:14.521013 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" podUID="3a5b66a7-1cf6-4840-870c-f7a6968bd08a" Nov 24 00:42:14 crc kubenswrapper[4888]: E1124 00:42:14.521099 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" podUID="82f2609b-e676-49f8-9ab4-d06fb3e626be" Nov 24 00:42:14 crc kubenswrapper[4888]: E1124 00:42:14.521096 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.146:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" podUID="bc56a88d-e719-40f5-a210-1b7afdffae8b" Nov 24 00:42:21 crc kubenswrapper[4888]: I1124 00:42:21.406347 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6d84888d56-jk5d2" Nov 24 00:42:23 crc kubenswrapper[4888]: I1124 00:42:23.591686 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:42:23 crc kubenswrapper[4888]: I1124 00:42:23.592260 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:42:24 crc kubenswrapper[4888]: E1124 00:42:24.286162 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96" Nov 24 00:42:24 crc kubenswrapper[4888]: E1124 00:42:24.286969 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22nzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-56f54d6746-wbxrh_openstack-operators(6cf992a1-eac0-4269-b8bb-e14554116727): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:42:30 crc kubenswrapper[4888]: E1124 00:42:30.004604 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 24 00:42:30 crc kubenswrapper[4888]: E1124 00:42:30.006287 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6dc6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-6mmtv_openstack-operators(c36e3742-c48a-48d1-96c8-87c3809a4183): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:42:30 crc kubenswrapper[4888]: E1124 00:42:30.548200 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b" Nov 24 00:42:30 crc kubenswrapper[4888]: E1124 00:42:30.548367 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-587bs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-c9cfn_openstack-operators(8a57fb36-c0db-4778-9c17-7241b5e8f602): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:42:31 crc kubenswrapper[4888]: E1124 00:42:31.058673 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04" Nov 24 00:42:31 crc kubenswrapper[4888]: E1124 00:42:31.058917 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dxjf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-54b5986bb8-4pmdz_openstack-operators(dbaba630-daf7-46c3-807a-6e76128fc7d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:42:31 crc kubenswrapper[4888]: E1124 00:42:31.527789 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894" Nov 24 00:42:31 crc kubenswrapper[4888]: E1124 00:42:31.528027 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5nspx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-6dd8864d7c-zl954_openstack-operators(e6420e55-4951-4546-b252-44c25d49a0c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:42:35 crc kubenswrapper[4888]: E1124 00:42:35.605772 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" podUID="6cf992a1-eac0-4269-b8bb-e14554116727" Nov 24 00:42:35 crc kubenswrapper[4888]: I1124 00:42:35.737602 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" event={"ID":"6cf992a1-eac0-4269-b8bb-e14554116727","Type":"ContainerStarted","Data":"76575eebcef1b21f13b348b5c01944b5a1d5decb3041c3539bdc9f28de597063"} Nov 24 00:42:35 crc kubenswrapper[4888]: E1124 00:42:35.829081 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" podUID="c36e3742-c48a-48d1-96c8-87c3809a4183" Nov 24 00:42:35 crc kubenswrapper[4888]: E1124 00:42:35.862460 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" podUID="8a57fb36-c0db-4778-9c17-7241b5e8f602" Nov 24 00:42:35 crc kubenswrapper[4888]: E1124 00:42:35.956854 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" podUID="dbaba630-daf7-46c3-807a-6e76128fc7d6" Nov 24 00:42:36 crc kubenswrapper[4888]: E1124 00:42:36.138486 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" podUID="e6420e55-4951-4546-b252-44c25d49a0c3" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.759543 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" event={"ID":"c36e3742-c48a-48d1-96c8-87c3809a4183","Type":"ContainerStarted","Data":"a16e9406bfe3ffa88cfebe6f60cf364bded3d6c92a201cad433125653197ce40"} Nov 24 00:42:36 crc kubenswrapper[4888]: E1124 00:42:36.763514 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" podUID="c36e3742-c48a-48d1-96c8-87c3809a4183" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.766879 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" event={"ID":"f4bc41bd-b9d5-4acf-978e-75144d8b43a3","Type":"ContainerStarted","Data":"7b0f1571d6b66bf4b6cec61255f26eed3a0f70276be87edd396cf32e5c981b1b"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.782292 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" event={"ID":"a277e2cc-4990-4d79-8a4f-ec03c81d747d","Type":"ContainerStarted","Data":"aaf70295b307146f51303c88efe330d2b6a748b8db359c3987a61fd9418ad64c"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.843438 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" event={"ID":"dbaba630-daf7-46c3-807a-6e76128fc7d6","Type":"ContainerStarted","Data":"b391d827f00844e588b5bc48158d35edfe6990880596f7e5b54c5efdb650fddb"} Nov 24 00:42:36 crc kubenswrapper[4888]: E1124 00:42:36.845265 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" podUID="dbaba630-daf7-46c3-807a-6e76128fc7d6" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.873761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" event={"ID":"afa14c5d-d820-4b05-bddd-4d00e49a421a","Type":"ContainerStarted","Data":"b83079fbc6c0c1138d99e684c757b9c9f969a4c81807d419dc369eadd576db1b"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.885415 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" event={"ID":"bc56a88d-e719-40f5-a210-1b7afdffae8b","Type":"ContainerStarted","Data":"53320d3acc2e1fab3fdeab16a3f163c4295f54cb90c6730b66702cb3b97d51b2"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.886183 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.887412 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" event={"ID":"8a57fb36-c0db-4778-9c17-7241b5e8f602","Type":"ContainerStarted","Data":"b0541ed0cc73244e7415ba7760bfa072024a4bd49ed8f15bdb20a663223c5832"} Nov 24 00:42:36 crc kubenswrapper[4888]: E1124 00:42:36.902866 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" podUID="8a57fb36-c0db-4778-9c17-7241b5e8f602" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.910240 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" event={"ID":"3a5b66a7-1cf6-4840-870c-f7a6968bd08a","Type":"ContainerStarted","Data":"2df8da94030c2e49a2fc74665d01ab98512f88c413a60e30e98e8ba339e33eef"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.911018 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.931220 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" event={"ID":"c18797ad-a3dc-46af-9aaf-d9e17918228a","Type":"ContainerStarted","Data":"b7e7773b39b1a31f9552fc625a2def7fd8b289a8bfc93c50ce1b9ee6c90a72b3"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.993345 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" event={"ID":"82f2609b-e676-49f8-9ab4-d06fb3e626be","Type":"ContainerStarted","Data":"d56bca0e9dfb6779bb6d7420b4df389be8ad2b3bbb1473052abafff7777b18e2"} Nov 24 00:42:36 crc kubenswrapper[4888]: I1124 00:42:36.993624 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.001834 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" event={"ID":"e6420e55-4951-4546-b252-44c25d49a0c3","Type":"ContainerStarted","Data":"72df9aaa48e6bfb171a5a15af4c10ca38bcf90b479aac7f028b596122796f705"} Nov 24 00:42:37 crc kubenswrapper[4888]: E1124 00:42:37.023084 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" podUID="e6420e55-4951-4546-b252-44c25d49a0c3" Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.026408 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" event={"ID":"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb","Type":"ContainerStarted","Data":"6a567036506bd4cc8973e2bf7f0b4ef9d6fa68b5833a20d9afeb67412cc7060e"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.040152 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" event={"ID":"10a23051-0913-44fd-abda-89436a7e3b59","Type":"ContainerStarted","Data":"b460b50ca3e08ad7a40e35ab72729a0d9f839ec1d5961a7f2fda1eab8cff3f20"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.052349 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" podStartSLOduration=4.513765208 podStartE2EDuration="28.052331177s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.93764431 +0000 UTC m=+1034.520328354" lastFinishedPulling="2025-11-24 00:42:35.476210269 +0000 UTC m=+1058.058894323" observedRunningTime="2025-11-24 00:42:37.001001882 +0000 UTC m=+1059.583685926" watchObservedRunningTime="2025-11-24 00:42:37.052331177 +0000 UTC m=+1059.635015221" Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.058521 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" event={"ID":"587fee04-3081-4c59-8125-0e47436e2347","Type":"ContainerStarted","Data":"38d27eb5ef53151eaf7de91be781dd0941e444602cf7c1d0a12c333bca9e36d8"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.061185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" event={"ID":"5fe80165-08e7-4a34-8290-fd0d7fd22dd2","Type":"ContainerStarted","Data":"c3a49131f254566ccf99de5f6fcb7ae710be8236f61259c0f4d9430dcb0bc7de"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.064723 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" event={"ID":"3766dfec-2de1-4df2-8609-0cbbe8599342","Type":"ContainerStarted","Data":"baf3d925640bec5253f268d0dd796e2ca914a445b74b7a3bb8aefe174ea73792"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.067254 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" event={"ID":"fef147fd-ee85-4c05-9652-5e44223a3e5e","Type":"ContainerStarted","Data":"5c8dd60aac4f1e97bbe42c4b6935ca9036e1ba0ced786423095ce7a172fd1259"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.068510 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" event={"ID":"a701224c-a2ad-49b6-98b7-55fd37051e5b","Type":"ContainerStarted","Data":"f00675420fae2dd5b70ad270ff3825d832480650f23b39c7f608aa4d4613c1b3"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.070026 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" event={"ID":"750c449f-a775-4ee8-82b8-a6be531250fe","Type":"ContainerStarted","Data":"d46058537bc86f1b3b568c5b5be64609b8d5707261f19a69c44089a88b237e5c"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.070858 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.088641 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" event={"ID":"e7342ffc-1f02-4eca-9be9-5816f5f0c86c","Type":"ContainerStarted","Data":"fa8ff33d8397c34047d07f3b29993343a96865313409470a73c47f1e09f0f66d"} Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.129816 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" podStartSLOduration=4.5628145700000005 podStartE2EDuration="28.129787663s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.962175586 +0000 UTC m=+1034.544859630" lastFinishedPulling="2025-11-24 00:42:35.529148669 +0000 UTC m=+1058.111832723" observedRunningTime="2025-11-24 00:42:37.123166178 +0000 UTC m=+1059.705850222" watchObservedRunningTime="2025-11-24 00:42:37.129787663 +0000 UTC m=+1059.712471707" Nov 24 00:42:37 crc kubenswrapper[4888]: I1124 00:42:37.224786 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" podStartSLOduration=4.68585485 podStartE2EDuration="28.224762709s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.935197811 +0000 UTC m=+1034.517881855" lastFinishedPulling="2025-11-24 00:42:35.47410567 +0000 UTC m=+1058.056789714" observedRunningTime="2025-11-24 00:42:37.201337594 +0000 UTC m=+1059.784021638" watchObservedRunningTime="2025-11-24 00:42:37.224762709 +0000 UTC m=+1059.807446753" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.097978 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" event={"ID":"e569ce1f-9272-4c6a-bb91-a01616bb5883","Type":"ContainerStarted","Data":"58e34c7bb18df56db8d4e7f8314ed2e1482987a8b30418a57971d0b7b384d24b"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.100089 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" event={"ID":"c18797ad-a3dc-46af-9aaf-d9e17918228a","Type":"ContainerStarted","Data":"4c1846fb87cb861a44ea380c911dc8757f20ff69f7ba9634df9215db02d2ffc6"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.100175 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.103067 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" event={"ID":"f4bc41bd-b9d5-4acf-978e-75144d8b43a3","Type":"ContainerStarted","Data":"7eaf4be4aaa6eabdb8d5dcf913f38e28ab12298f72b56745e7b8f053c597005b"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.103186 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.105154 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" event={"ID":"10a23051-0913-44fd-abda-89436a7e3b59","Type":"ContainerStarted","Data":"022b1d783fe6ebbafbddb4547d590762cf6c5c3da68df404c856da8f7b8bb6cf"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.105248 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.107595 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" event={"ID":"3766dfec-2de1-4df2-8609-0cbbe8599342","Type":"ContainerStarted","Data":"492c793da10752cac4070800ced0eb5ab69eb3fc1a7ed6f2f9f1a4521e908b8d"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.107753 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.110024 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" event={"ID":"2b35fb96-9ea1-4aaa-9396-cd5f705f63bb","Type":"ContainerStarted","Data":"551334954b5aa3216001598a5fd223f8b5c19d86223f3ef2529b368aea74720e"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.110111 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.112688 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" event={"ID":"5fe80165-08e7-4a34-8290-fd0d7fd22dd2","Type":"ContainerStarted","Data":"5a7ec34fe8675c1109183cc3e7e8a1c42bfacbdf166ddabf2ab0810c311c53b9"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.112791 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.114337 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" event={"ID":"e7342ffc-1f02-4eca-9be9-5816f5f0c86c","Type":"ContainerStarted","Data":"380ae1286bb7f6c3f1e00ad59213c6944d36c4d4a6307233586d3b86152ab7fb"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.115259 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.117317 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" event={"ID":"fef147fd-ee85-4c05-9652-5e44223a3e5e","Type":"ContainerStarted","Data":"340186bb4e889ade876765c05f96e215b368da5f3ff1723654ba5fa42b789054"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.117988 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.122345 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" event={"ID":"a701224c-a2ad-49b6-98b7-55fd37051e5b","Type":"ContainerStarted","Data":"658b86c52e0e87f3b8a1f16f6174addab41ce0c84abb318b02f298f7aee6f50d"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.123342 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.123791 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv" podStartSLOduration=4.462103019 podStartE2EDuration="28.123776231s" podCreationTimestamp="2025-11-24 00:42:10 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.930652654 +0000 UTC m=+1034.513336698" lastFinishedPulling="2025-11-24 00:42:35.592325856 +0000 UTC m=+1058.175009910" observedRunningTime="2025-11-24 00:42:38.116741354 +0000 UTC m=+1060.699425408" watchObservedRunningTime="2025-11-24 00:42:38.123776231 +0000 UTC m=+1060.706460275" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.124091 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" podStartSLOduration=5.611735694 podStartE2EDuration="29.12408578s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.962283709 +0000 UTC m=+1034.544967753" lastFinishedPulling="2025-11-24 00:42:35.474633775 +0000 UTC m=+1058.057317839" observedRunningTime="2025-11-24 00:42:37.310029054 +0000 UTC m=+1059.892713098" watchObservedRunningTime="2025-11-24 00:42:38.12408578 +0000 UTC m=+1060.706769824" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.125479 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" event={"ID":"6cf992a1-eac0-4269-b8bb-e14554116727","Type":"ContainerStarted","Data":"16ecd626a4c5cc843d084926b69143f51d79e830a29bb090592a8c81d00209ff"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.125628 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.130196 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" event={"ID":"a277e2cc-4990-4d79-8a4f-ec03c81d747d","Type":"ContainerStarted","Data":"c445e6439a8cfa74655ef1752950d6d021674bc2aa86daa98484a6e38ff99984"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.131201 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.135290 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" event={"ID":"afa14c5d-d820-4b05-bddd-4d00e49a421a","Type":"ContainerStarted","Data":"c6db97537fd2c98c6a303e864f90318187df5f3434ba183c3f20ded3f0f64022"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.136244 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.139304 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" event={"ID":"587fee04-3081-4c59-8125-0e47436e2347","Type":"ContainerStarted","Data":"1f7432f3385ea6406e333abc8174609b80ab427f2776a05d0da0348e1465d7ab"} Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.139349 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:38 crc kubenswrapper[4888]: E1124 00:42:38.142437 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" podUID="c36e3742-c48a-48d1-96c8-87c3809a4183" Nov 24 00:42:38 crc kubenswrapper[4888]: E1124 00:42:38.143564 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" podUID="e6420e55-4951-4546-b252-44c25d49a0c3" Nov 24 00:42:38 crc kubenswrapper[4888]: E1124 00:42:38.143565 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" podUID="dbaba630-daf7-46c3-807a-6e76128fc7d6" Nov 24 00:42:38 crc kubenswrapper[4888]: E1124 00:42:38.150371 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" podUID="8a57fb36-c0db-4778-9c17-7241b5e8f602" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.160679 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" podStartSLOduration=8.548701451 podStartE2EDuration="29.160665043s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:12.347700188 +0000 UTC m=+1034.930384232" lastFinishedPulling="2025-11-24 00:42:32.95966378 +0000 UTC m=+1055.542347824" observedRunningTime="2025-11-24 00:42:38.158840632 +0000 UTC m=+1060.741524676" watchObservedRunningTime="2025-11-24 00:42:38.160665043 +0000 UTC m=+1060.743349087" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.180855 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" podStartSLOduration=7.672942678 podStartE2EDuration="29.180838307s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.453129519 +0000 UTC m=+1034.035813563" lastFinishedPulling="2025-11-24 00:42:32.961025138 +0000 UTC m=+1055.543709192" observedRunningTime="2025-11-24 00:42:38.180326342 +0000 UTC m=+1060.763010386" watchObservedRunningTime="2025-11-24 00:42:38.180838307 +0000 UTC m=+1060.763522351" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.227682 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" podStartSLOduration=7.4445613999999996 podStartE2EDuration="29.227667936s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.179068045 +0000 UTC m=+1033.761752089" lastFinishedPulling="2025-11-24 00:42:32.962174581 +0000 UTC m=+1055.544858625" observedRunningTime="2025-11-24 00:42:38.208008257 +0000 UTC m=+1060.790692301" watchObservedRunningTime="2025-11-24 00:42:38.227667936 +0000 UTC m=+1060.810351980" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.232389 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" podStartSLOduration=6.402434586 podStartE2EDuration="29.232383158s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.445005362 +0000 UTC m=+1034.027689406" lastFinishedPulling="2025-11-24 00:42:34.274953934 +0000 UTC m=+1056.857637978" observedRunningTime="2025-11-24 00:42:38.225329861 +0000 UTC m=+1060.808013905" watchObservedRunningTime="2025-11-24 00:42:38.232383158 +0000 UTC m=+1060.815067202" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.247098 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" podStartSLOduration=6.150641715 podStartE2EDuration="29.247082439s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.1785421 +0000 UTC m=+1033.761226144" lastFinishedPulling="2025-11-24 00:42:34.274982814 +0000 UTC m=+1056.857666868" observedRunningTime="2025-11-24 00:42:38.240724282 +0000 UTC m=+1060.823408326" watchObservedRunningTime="2025-11-24 00:42:38.247082439 +0000 UTC m=+1060.829766483" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.265545 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" podStartSLOduration=7.160363702 podStartE2EDuration="29.265527855s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:10.857345907 +0000 UTC m=+1033.440029951" lastFinishedPulling="2025-11-24 00:42:32.96251006 +0000 UTC m=+1055.545194104" observedRunningTime="2025-11-24 00:42:38.262517621 +0000 UTC m=+1060.845201665" watchObservedRunningTime="2025-11-24 00:42:38.265527855 +0000 UTC m=+1060.848211889" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.283080 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" podStartSLOduration=7.113646547 podStartE2EDuration="29.283063886s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:10.791572568 +0000 UTC m=+1033.374256622" lastFinishedPulling="2025-11-24 00:42:32.960989917 +0000 UTC m=+1055.543673961" observedRunningTime="2025-11-24 00:42:38.279769264 +0000 UTC m=+1060.862453308" watchObservedRunningTime="2025-11-24 00:42:38.283063886 +0000 UTC m=+1060.865747940" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.301077 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" podStartSLOduration=7.789030885 podStartE2EDuration="29.301052179s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.450130986 +0000 UTC m=+1034.032815030" lastFinishedPulling="2025-11-24 00:42:32.96215229 +0000 UTC m=+1055.544836324" observedRunningTime="2025-11-24 00:42:38.295845143 +0000 UTC m=+1060.878529187" watchObservedRunningTime="2025-11-24 00:42:38.301052179 +0000 UTC m=+1060.883736243" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.311492 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" podStartSLOduration=4.240960848 podStartE2EDuration="29.31147051s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.184853567 +0000 UTC m=+1033.767537601" lastFinishedPulling="2025-11-24 00:42:36.255363219 +0000 UTC m=+1058.838047263" observedRunningTime="2025-11-24 00:42:38.309769723 +0000 UTC m=+1060.892453767" watchObservedRunningTime="2025-11-24 00:42:38.31147051 +0000 UTC m=+1060.894154554" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.375779 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" podStartSLOduration=8.028060599 podStartE2EDuration="29.375752468s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.615069488 +0000 UTC m=+1034.197753532" lastFinishedPulling="2025-11-24 00:42:32.962761357 +0000 UTC m=+1055.545445401" observedRunningTime="2025-11-24 00:42:38.371382776 +0000 UTC m=+1060.954066820" watchObservedRunningTime="2025-11-24 00:42:38.375752468 +0000 UTC m=+1060.958436512" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.403785 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" podStartSLOduration=8.359872428 podStartE2EDuration="29.403771661s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.916640302 +0000 UTC m=+1034.499324346" lastFinishedPulling="2025-11-24 00:42:32.960539535 +0000 UTC m=+1055.543223579" observedRunningTime="2025-11-24 00:42:38.401188529 +0000 UTC m=+1060.983872563" watchObservedRunningTime="2025-11-24 00:42:38.403771661 +0000 UTC m=+1060.986455705" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.421565 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" podStartSLOduration=7.415993511 podStartE2EDuration="29.421544308s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:10.956327185 +0000 UTC m=+1033.539011229" lastFinishedPulling="2025-11-24 00:42:32.961877972 +0000 UTC m=+1055.544562026" observedRunningTime="2025-11-24 00:42:38.415054147 +0000 UTC m=+1060.997738201" watchObservedRunningTime="2025-11-24 00:42:38.421544308 +0000 UTC m=+1061.004228362" Nov 24 00:42:38 crc kubenswrapper[4888]: I1124 00:42:38.432526 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" podStartSLOduration=7.329544914 podStartE2EDuration="29.432511865s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:10.858643314 +0000 UTC m=+1033.441327358" lastFinishedPulling="2025-11-24 00:42:32.961610265 +0000 UTC m=+1055.544294309" observedRunningTime="2025-11-24 00:42:38.427459804 +0000 UTC m=+1061.010143848" watchObservedRunningTime="2025-11-24 00:42:38.432511865 +0000 UTC m=+1061.015195909" Nov 24 00:42:40 crc kubenswrapper[4888]: I1124 00:42:40.028178 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-tsb9j" Nov 24 00:42:40 crc kubenswrapper[4888]: I1124 00:42:40.678353 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-mn678" Nov 24 00:42:41 crc kubenswrapper[4888]: I1124 00:42:41.709035 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.254154 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" event={"ID":"c36e3742-c48a-48d1-96c8-87c3809a4183","Type":"ContainerStarted","Data":"4df108de0cf6c4d0a7352c9737d76b68a3e35f7fba9c6e640db25f7b563ba3bc"} Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.254904 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.275765 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" podStartSLOduration=3.658205941 podStartE2EDuration="40.2757436s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:12.083078887 +0000 UTC m=+1034.665762931" lastFinishedPulling="2025-11-24 00:42:48.700616546 +0000 UTC m=+1071.283300590" observedRunningTime="2025-11-24 00:42:49.274356841 +0000 UTC m=+1071.857040895" watchObservedRunningTime="2025-11-24 00:42:49.2757436 +0000 UTC m=+1071.858427654" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.547659 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-gphtr" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.566076 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-9nb4l" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.650293 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-4gh28" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.685605 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-mtqr5" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.712051 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-wbxrh" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.854862 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-llt2h" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.876320 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jk6ph" Nov 24 00:42:49 crc kubenswrapper[4888]: I1124 00:42:49.990350 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-8kqtd" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.057695 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-fn4cs" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.144515 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-7zjtv" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.172985 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-vxffn" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.263060 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-98b6n" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.265796 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" event={"ID":"e6420e55-4951-4546-b252-44c25d49a0c3","Type":"ContainerStarted","Data":"d051262aa5556ce28d6c5357f0f41b90bcc3c0a5194ee8ced39121ae1eabd173"} Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.265998 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.305161 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" podStartSLOduration=3.48828106 podStartE2EDuration="41.305144089s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.922908668 +0000 UTC m=+1034.505592712" lastFinishedPulling="2025-11-24 00:42:49.739771697 +0000 UTC m=+1072.322455741" observedRunningTime="2025-11-24 00:42:50.304503701 +0000 UTC m=+1072.887187755" watchObservedRunningTime="2025-11-24 00:42:50.305144089 +0000 UTC m=+1072.887828133" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.583753 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-5ck2h" Nov 24 00:42:50 crc kubenswrapper[4888]: I1124 00:42:50.645609 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67c4db9997-7rqlj" Nov 24 00:42:51 crc kubenswrapper[4888]: I1124 00:42:51.275485 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" event={"ID":"dbaba630-daf7-46c3-807a-6e76128fc7d6","Type":"ContainerStarted","Data":"9a94d6b294b9fa39317cadad602c41e1ee3f5a0be2e8e1c95455a1895a28a421"} Nov 24 00:42:51 crc kubenswrapper[4888]: I1124 00:42:51.275752 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:42:51 crc kubenswrapper[4888]: I1124 00:42:51.299397 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" podStartSLOduration=3.2505456710000002 podStartE2EDuration="42.299380215s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.901220141 +0000 UTC m=+1034.483904185" lastFinishedPulling="2025-11-24 00:42:50.950054665 +0000 UTC m=+1073.532738729" observedRunningTime="2025-11-24 00:42:51.296292588 +0000 UTC m=+1073.878976642" watchObservedRunningTime="2025-11-24 00:42:51.299380215 +0000 UTC m=+1073.882064249" Nov 24 00:42:53 crc kubenswrapper[4888]: I1124 00:42:53.298168 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" event={"ID":"8a57fb36-c0db-4778-9c17-7241b5e8f602","Type":"ContainerStarted","Data":"29362fcece3ea53a895da8f26f5f5a45ecf7d9adf5167aa7c94a972c13391263"} Nov 24 00:42:53 crc kubenswrapper[4888]: I1124 00:42:53.298945 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:42:53 crc kubenswrapper[4888]: I1124 00:42:53.330167 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" podStartSLOduration=3.582859564 podStartE2EDuration="44.330140827s" podCreationTimestamp="2025-11-24 00:42:09 +0000 UTC" firstStartedPulling="2025-11-24 00:42:11.930365606 +0000 UTC m=+1034.513049640" lastFinishedPulling="2025-11-24 00:42:52.677646859 +0000 UTC m=+1075.260330903" observedRunningTime="2025-11-24 00:42:53.321402353 +0000 UTC m=+1075.904086437" watchObservedRunningTime="2025-11-24 00:42:53.330140827 +0000 UTC m=+1075.912824901" Nov 24 00:42:53 crc kubenswrapper[4888]: I1124 00:42:53.592309 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:42:53 crc kubenswrapper[4888]: I1124 00:42:53.592785 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:43:00 crc kubenswrapper[4888]: I1124 00:43:00.044939 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-4pmdz" Nov 24 00:43:00 crc kubenswrapper[4888]: I1124 00:43:00.263345 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-c9cfn" Nov 24 00:43:00 crc kubenswrapper[4888]: I1124 00:43:00.379499 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-zl954" Nov 24 00:43:00 crc kubenswrapper[4888]: I1124 00:43:00.753184 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-6mmtv" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.163087 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.166231 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.170646 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.171558 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.171935 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-j965k" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.172105 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.172272 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.222992 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stggl\" (UniqueName: \"kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.223039 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.256709 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.258069 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.258656 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.261064 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.324003 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stggl\" (UniqueName: \"kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.324285 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.325212 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.342725 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stggl\" (UniqueName: \"kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl\") pod \"dnsmasq-dns-675f4bcbfc-fqcm7\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.426070 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmwm2\" (UniqueName: \"kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.426245 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.426272 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.528187 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.528233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.528284 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmwm2\" (UniqueName: \"kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.529202 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.529256 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.542802 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.547600 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmwm2\" (UniqueName: \"kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2\") pod \"dnsmasq-dns-78dd6ddcc-z7j49\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.578257 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.591707 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.591780 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.591849 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.592525 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:43:23 crc kubenswrapper[4888]: I1124 00:43:23.592580 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5" gracePeriod=600 Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.009123 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:24 crc kubenswrapper[4888]: W1124 00:43:24.013793 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22d70f3e_96b4_4ea7_aad9_6d51b521731e.slice/crio-b689da6b06c87aac33eec626997870c7725ce92e36828a3a709f592c97b72ecd WatchSource:0}: Error finding container b689da6b06c87aac33eec626997870c7725ce92e36828a3a709f592c97b72ecd: Status 404 returned error can't find the container with id b689da6b06c87aac33eec626997870c7725ce92e36828a3a709f592c97b72ecd Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.074324 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:24 crc kubenswrapper[4888]: W1124 00:43:24.077052 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcce0305_17cb_479c_b8b5_13a0fe2a7830.slice/crio-fa543fe704278f1402819db54a1efdbba05681d5e4d6edef7c82783185668694 WatchSource:0}: Error finding container fa543fe704278f1402819db54a1efdbba05681d5e4d6edef7c82783185668694: Status 404 returned error can't find the container with id fa543fe704278f1402819db54a1efdbba05681d5e4d6edef7c82783185668694 Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.610520 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" event={"ID":"22d70f3e-96b4-4ea7-aad9-6d51b521731e","Type":"ContainerStarted","Data":"b689da6b06c87aac33eec626997870c7725ce92e36828a3a709f592c97b72ecd"} Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.616127 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5" exitCode=0 Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.616186 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5"} Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.616216 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c"} Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.616234 4888 scope.go:117] "RemoveContainer" containerID="972439db68e4b83dc064ccb0f06526e764d5f6b047d97d2fdfd6133f10f97778" Nov 24 00:43:24 crc kubenswrapper[4888]: I1124 00:43:24.619318 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" event={"ID":"fcce0305-17cb-479c-b8b5-13a0fe2a7830","Type":"ContainerStarted","Data":"fa543fe704278f1402819db54a1efdbba05681d5e4d6edef7c82783185668694"} Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.123873 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.153848 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.155585 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.205151 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.296553 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssh28\" (UniqueName: \"kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.296627 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.296750 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.399105 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssh28\" (UniqueName: \"kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.399174 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.399253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.400687 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.403201 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.431729 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.446546 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.448175 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.450069 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssh28\" (UniqueName: \"kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28\") pod \"dnsmasq-dns-666b6646f7-fzvnt\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.470112 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.495597 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.603364 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.603747 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtctx\" (UniqueName: \"kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.603787 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.705002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtctx\" (UniqueName: \"kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.705073 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.705137 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.706062 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.706792 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.752921 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtctx\" (UniqueName: \"kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx\") pod \"dnsmasq-dns-57d769cc4f-f8bps\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.779016 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:26 crc kubenswrapper[4888]: I1124 00:43:26.997518 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.237657 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.290544 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.292347 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296122 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296285 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-gqh27" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296355 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296285 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296723 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.296933 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.297308 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.308494 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422693 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422747 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc5dl\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422885 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422909 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422958 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.422988 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.423027 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.423049 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.423065 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.423081 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.524614 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.524938 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.524963 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525846 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525887 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525887 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525912 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc5dl\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525953 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525976 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.525999 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.526049 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.526115 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.526539 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.527188 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.538690 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.538835 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.538985 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.540726 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.541858 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.542308 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.544039 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.554710 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc5dl\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.564313 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.564563 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.577260 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.577376 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.579749 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.581666 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.581878 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.582114 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.582214 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lvpwf" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.582377 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.582418 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.623520 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.659869 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" event={"ID":"79b310c2-f53b-401e-89e1-60e7cec6d274","Type":"ContainerStarted","Data":"8847fef06b095a423bbbabc76e85bc7430348ae39f060f87cfac3a0ce4e08dbf"} Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729233 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729315 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729366 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729438 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729515 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729580 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9chg\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729639 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729717 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729750 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729798 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.729876 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831252 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831313 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831356 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831412 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831452 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831502 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831542 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831590 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831644 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9chg\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831671 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.831722 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.833241 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.835067 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.835365 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.835421 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.837337 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.837521 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.841725 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.842133 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.843272 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.858092 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9chg\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.860469 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.881744 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:27 crc kubenswrapper[4888]: I1124 00:43:27.921168 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.047412 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.050643 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.054225 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.055710 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-s8wnz" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.056622 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.058382 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.060756 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.069119 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159475 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159634 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159691 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159725 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqtsr\" (UniqueName: \"kubernetes.io/projected/d1c29616-6ecc-4f98-9849-b04242cf0a66-kube-api-access-pqtsr\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159773 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159877 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159937 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.159956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261132 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261176 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261215 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqtsr\" (UniqueName: \"kubernetes.io/projected/d1c29616-6ecc-4f98-9849-b04242cf0a66-kube-api-access-pqtsr\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261230 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261282 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261302 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261323 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261356 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.261629 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.264285 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.265335 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.269678 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.270115 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.270151 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1c29616-6ecc-4f98-9849-b04242cf0a66-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.273275 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c29616-6ecc-4f98-9849-b04242cf0a66-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.286409 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.288781 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqtsr\" (UniqueName: \"kubernetes.io/projected/d1c29616-6ecc-4f98-9849-b04242cf0a66-kube-api-access-pqtsr\") pod \"openstack-galera-0\" (UID: \"d1c29616-6ecc-4f98-9849-b04242cf0a66\") " pod="openstack/openstack-galera-0" Nov 24 00:43:29 crc kubenswrapper[4888]: I1124 00:43:29.373350 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: W1124 00:43:30.079392 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df49d3e_c85e_4969_adf0_9aee1f672f4a.slice/crio-a431b4effe092bb901a95e439c84bfbc3295b186274dde1b0b409df6a85f7ced WatchSource:0}: Error finding container a431b4effe092bb901a95e439c84bfbc3295b186274dde1b0b409df6a85f7ced: Status 404 returned error can't find the container with id a431b4effe092bb901a95e439c84bfbc3295b186274dde1b0b409df6a85f7ced Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.543515 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.545255 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.550306 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-x5nmw" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.550580 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.550767 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.551436 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.559430 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692286 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692339 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692388 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692428 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692489 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j8gq\" (UniqueName: \"kubernetes.io/projected/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kube-api-access-4j8gq\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692526 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692560 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.692603 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.704125 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" event={"ID":"6df49d3e-c85e-4969-adf0-9aee1f672f4a","Type":"ContainerStarted","Data":"a431b4effe092bb901a95e439c84bfbc3295b186274dde1b0b409df6a85f7ced"} Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.790628 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.791697 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.793933 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794249 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794300 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794348 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794387 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794442 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j8gq\" (UniqueName: \"kubernetes.io/projected/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kube-api-access-4j8gq\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794479 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794517 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.794562 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.795222 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.795403 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.795598 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dvhhv" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.798753 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.798980 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.800004 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.805422 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.808770 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.811782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.812345 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.833431 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.841114 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j8gq\" (UniqueName: \"kubernetes.io/projected/9a963d44-99cd-45d4-afd8-9c9c7bfb95e7-kube-api-access-4j8gq\") pod \"openstack-cell1-galera-0\" (UID: \"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7\") " pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.882792 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.897656 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.897739 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.897770 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kolla-config\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.897792 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2bbc\" (UniqueName: \"kubernetes.io/projected/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kube-api-access-f2bbc\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:30 crc kubenswrapper[4888]: I1124 00:43:30.897857 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-config-data\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.000507 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-config-data\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.000635 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.000703 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.000725 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kolla-config\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.000750 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2bbc\" (UniqueName: \"kubernetes.io/projected/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kube-api-access-f2bbc\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.001549 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-config-data\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.002208 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kolla-config\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.004381 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.007697 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fca78ec3-fec9-4c3e-aee9-75d3e049c668-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.025134 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2bbc\" (UniqueName: \"kubernetes.io/projected/fca78ec3-fec9-4c3e-aee9-75d3e049c668-kube-api-access-f2bbc\") pod \"memcached-0\" (UID: \"fca78ec3-fec9-4c3e-aee9-75d3e049c668\") " pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.225269 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 00:43:31 crc kubenswrapper[4888]: I1124 00:43:31.434229 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.740485 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.741754 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.744141 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-snmhh" Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.754897 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.841910 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28kxf\" (UniqueName: \"kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf\") pod \"kube-state-metrics-0\" (UID: \"4678859e-43e2-4d0b-ae68-323ec7239ad7\") " pod="openstack/kube-state-metrics-0" Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.943701 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28kxf\" (UniqueName: \"kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf\") pod \"kube-state-metrics-0\" (UID: \"4678859e-43e2-4d0b-ae68-323ec7239ad7\") " pod="openstack/kube-state-metrics-0" Nov 24 00:43:32 crc kubenswrapper[4888]: I1124 00:43:32.968014 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28kxf\" (UniqueName: \"kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf\") pod \"kube-state-metrics-0\" (UID: \"4678859e-43e2-4d0b-ae68-323ec7239ad7\") " pod="openstack/kube-state-metrics-0" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.064433 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.416585 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk"] Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.417795 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.420701 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.420976 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-9d7ks" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.432150 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk"] Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.555100 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj4c6\" (UniqueName: \"kubernetes.io/projected/a05abc8e-614d-4cb7-8eda-35f0698a756c-kube-api-access-hj4c6\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.555438 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05abc8e-614d-4cb7-8eda-35f0698a756c-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.657404 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj4c6\" (UniqueName: \"kubernetes.io/projected/a05abc8e-614d-4cb7-8eda-35f0698a756c-kube-api-access-hj4c6\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.657468 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05abc8e-614d-4cb7-8eda-35f0698a756c-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.673622 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05abc8e-614d-4cb7-8eda-35f0698a756c-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.687534 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj4c6\" (UniqueName: \"kubernetes.io/projected/a05abc8e-614d-4cb7-8eda-35f0698a756c-kube-api-access-hj4c6\") pod \"observability-ui-dashboards-7d5fb4cbfb-clzxk\" (UID: \"a05abc8e-614d-4cb7-8eda-35f0698a756c\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.731956 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5dcb56c6bb-bgkr9"] Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.733170 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.738795 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.742543 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dcb56c6bb-bgkr9"] Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861113 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-oauth-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861362 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-trusted-ca-bundle\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861447 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861545 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-console-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861620 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-oauth-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861678 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-service-ca\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.861715 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plcl6\" (UniqueName: \"kubernetes.io/projected/8307845f-a311-4346-8012-516220cf33b9-kube-api-access-plcl6\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962735 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-console-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962787 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-oauth-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962813 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-service-ca\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962847 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plcl6\" (UniqueName: \"kubernetes.io/projected/8307845f-a311-4346-8012-516220cf33b9-kube-api-access-plcl6\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962910 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-oauth-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962966 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-trusted-ca-bundle\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.962991 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.963697 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-console-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.963721 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-oauth-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.963741 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-service-ca\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.964393 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8307845f-a311-4346-8012-516220cf33b9-trusted-ca-bundle\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.966532 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-serving-cert\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.977915 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8307845f-a311-4346-8012-516220cf33b9-console-oauth-config\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:33 crc kubenswrapper[4888]: I1124 00:43:33.990773 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plcl6\" (UniqueName: \"kubernetes.io/projected/8307845f-a311-4346-8012-516220cf33b9-kube-api-access-plcl6\") pod \"console-5dcb56c6bb-bgkr9\" (UID: \"8307845f-a311-4346-8012-516220cf33b9\") " pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.051452 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.060803 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.063177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.066540 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.066678 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.067142 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.067164 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.067297 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-5vsrn" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.075765 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.079494 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165515 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165582 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165607 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165730 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165857 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv7jq\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165893 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.165938 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.166011 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269431 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv7jq\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269499 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269551 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269600 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269634 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269681 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269708 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.269760 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.270501 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.273219 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.273551 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.274459 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.274486 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/94ad0ac163fed009bb5498203f4a81a25a208cff03033a2bd0264ed6773db54c/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.274514 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.274782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.277033 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.289437 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv7jq\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.305234 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:34 crc kubenswrapper[4888]: I1124 00:43:34.378062 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:43:35 crc kubenswrapper[4888]: I1124 00:43:35.759220 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1c29616-6ecc-4f98-9849-b04242cf0a66","Type":"ContainerStarted","Data":"1a65bb63febde54996ba5d1d3b6365c20264620520c42dd458e885527b175946"} Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.440506 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x5tvh"] Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.441926 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.447134 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-5j4q6" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.448347 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.448490 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.463053 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x5tvh"] Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.509940 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-g6h6d"] Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.522173 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.549640 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-g6h6d"] Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564764 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-ovn-controller-tls-certs\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564810 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564856 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-combined-ca-bundle\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564908 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvzdz\" (UniqueName: \"kubernetes.io/projected/053aa355-b781-49f4-8e8a-80ae64c4de62-kube-api-access-vvzdz\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564954 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-log-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.564980 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053aa355-b781-49f4-8e8a-80ae64c4de62-scripts\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.565114 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667192 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-log\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667237 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667265 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-ovn-controller-tls-certs\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667318 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb427d6c-c0ae-454c-8c53-71d20209ea47-scripts\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667376 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-combined-ca-bundle\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667507 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvzdz\" (UniqueName: \"kubernetes.io/projected/053aa355-b781-49f4-8e8a-80ae64c4de62-kube-api-access-vvzdz\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667553 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwk6\" (UniqueName: \"kubernetes.io/projected/fb427d6c-c0ae-454c-8c53-71d20209ea47-kube-api-access-bdwk6\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667623 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-etc-ovs\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667643 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-log-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053aa355-b781-49f4-8e8a-80ae64c4de62-scripts\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667769 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-lib\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667835 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667885 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-run\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.667905 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.668137 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-log-ovn\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.668219 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/053aa355-b781-49f4-8e8a-80ae64c4de62-var-run\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.671426 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053aa355-b781-49f4-8e8a-80ae64c4de62-scripts\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.686468 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-combined-ca-bundle\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.694752 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/053aa355-b781-49f4-8e8a-80ae64c4de62-ovn-controller-tls-certs\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.704426 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvzdz\" (UniqueName: \"kubernetes.io/projected/053aa355-b781-49f4-8e8a-80ae64c4de62-kube-api-access-vvzdz\") pod \"ovn-controller-x5tvh\" (UID: \"053aa355-b781-49f4-8e8a-80ae64c4de62\") " pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.762607 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769204 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwk6\" (UniqueName: \"kubernetes.io/projected/fb427d6c-c0ae-454c-8c53-71d20209ea47-kube-api-access-bdwk6\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769320 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-etc-ovs\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769387 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-lib\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769438 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-run\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769501 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-log\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.769525 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb427d6c-c0ae-454c-8c53-71d20209ea47-scripts\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.770142 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-lib\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.770309 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-etc-ovs\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.770370 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-run\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.770470 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb427d6c-c0ae-454c-8c53-71d20209ea47-var-log\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.771626 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb427d6c-c0ae-454c-8c53-71d20209ea47-scripts\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.790807 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwk6\" (UniqueName: \"kubernetes.io/projected/fb427d6c-c0ae-454c-8c53-71d20209ea47-kube-api-access-bdwk6\") pod \"ovn-controller-ovs-g6h6d\" (UID: \"fb427d6c-c0ae-454c-8c53-71d20209ea47\") " pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:37 crc kubenswrapper[4888]: I1124 00:43:37.847322 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.292331 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.294366 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.297456 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.297516 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.297456 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.297876 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.297896 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dp58q" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.320948 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378715 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378756 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cdzf\" (UniqueName: \"kubernetes.io/projected/728a0118-33ee-4fb3-b8e2-198e075adc31-kube-api-access-5cdzf\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378791 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378889 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378943 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378961 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.378997 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-config\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.379034 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480633 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480678 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cdzf\" (UniqueName: \"kubernetes.io/projected/728a0118-33ee-4fb3-b8e2-198e075adc31-kube-api-access-5cdzf\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480701 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480742 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480770 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480786 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480802 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-config\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.480845 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.481585 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.481769 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.483091 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.482770 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/728a0118-33ee-4fb3-b8e2-198e075adc31-config\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.484844 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.486495 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.495094 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728a0118-33ee-4fb3-b8e2-198e075adc31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.497068 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cdzf\" (UniqueName: \"kubernetes.io/projected/728a0118-33ee-4fb3-b8e2-198e075adc31-kube-api-access-5cdzf\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.513654 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"728a0118-33ee-4fb3-b8e2-198e075adc31\") " pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:38 crc kubenswrapper[4888]: I1124 00:43:38.621834 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.149973 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.151630 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.154199 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7nwht" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.154247 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.155078 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.157298 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.169472 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.214969 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215050 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-config\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215073 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215108 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pxj8\" (UniqueName: \"kubernetes.io/projected/cdce96a2-6d42-4761-9170-ef61211cfc92-kube-api-access-8pxj8\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215151 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215168 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215190 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.215228 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316457 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-config\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316599 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316661 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pxj8\" (UniqueName: \"kubernetes.io/projected/cdce96a2-6d42-4761-9170-ef61211cfc92-kube-api-access-8pxj8\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316741 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316774 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316844 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.316906 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.323738 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.324255 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.324378 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-config\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.324521 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdce96a2-6d42-4761-9170-ef61211cfc92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.330907 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.331622 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.351189 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdce96a2-6d42-4761-9170-ef61211cfc92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.351458 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pxj8\" (UniqueName: \"kubernetes.io/projected/cdce96a2-6d42-4761-9170-ef61211cfc92-kube-api-access-8pxj8\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.353201 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cdce96a2-6d42-4761-9170-ef61211cfc92\") " pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:40 crc kubenswrapper[4888]: I1124 00:43:40.496154 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 00:43:41 crc kubenswrapper[4888]: I1124 00:43:41.135748 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 00:43:41 crc kubenswrapper[4888]: I1124 00:43:41.143929 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:43:41 crc kubenswrapper[4888]: I1124 00:43:41.246194 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:43:44 crc kubenswrapper[4888]: W1124 00:43:44.547966 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a963d44_99cd_45d4_afd8_9c9c7bfb95e7.slice/crio-49fc229a4ce0e2da3e81845d46eded454d092890426e614cd9a91450bfe2b5df WatchSource:0}: Error finding container 49fc229a4ce0e2da3e81845d46eded454d092890426e614cd9a91450bfe2b5df: Status 404 returned error can't find the container with id 49fc229a4ce0e2da3e81845d46eded454d092890426e614cd9a91450bfe2b5df Nov 24 00:43:44 crc kubenswrapper[4888]: W1124 00:43:44.561829 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc51b1de1_34e6_48a3_8950_cdfaf80d28a0.slice/crio-8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42 WatchSource:0}: Error finding container 8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42: Status 404 returned error can't find the container with id 8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42 Nov 24 00:43:44 crc kubenswrapper[4888]: W1124 00:43:44.567115 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19077405_4b37_4747_a4aa_37013176475e.slice/crio-60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf WatchSource:0}: Error finding container 60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf: Status 404 returned error can't find the container with id 60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.606510 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.606715 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stggl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-fqcm7_openstack(22d70f3e-96b4-4ea7-aad9-6d51b521731e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.608084 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" podUID="22d70f3e-96b4-4ea7-aad9-6d51b521731e" Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.645012 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.645237 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pmwm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-z7j49_openstack(fcce0305-17cb-479c-b8b5-13a0fe2a7830): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:43:44 crc kubenswrapper[4888]: E1124 00:43:44.646776 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" podUID="fcce0305-17cb-479c-b8b5-13a0fe2a7830" Nov 24 00:43:44 crc kubenswrapper[4888]: I1124 00:43:44.874953 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7","Type":"ContainerStarted","Data":"49fc229a4ce0e2da3e81845d46eded454d092890426e614cd9a91450bfe2b5df"} Nov 24 00:43:44 crc kubenswrapper[4888]: I1124 00:43:44.880100 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerStarted","Data":"8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42"} Nov 24 00:43:44 crc kubenswrapper[4888]: I1124 00:43:44.882942 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerStarted","Data":"60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.060935 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.072620 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dcb56c6bb-bgkr9"] Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.424569 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:43:45 crc kubenswrapper[4888]: W1124 00:43:45.430515 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce885b04_03a7_43c9_acd0_f26239bf4596.slice/crio-5b03db74d02b832cd0155e69110494340500923a78bca550c596feed69266d92 WatchSource:0}: Error finding container 5b03db74d02b832cd0155e69110494340500923a78bca550c596feed69266d92: Status 404 returned error can't find the container with id 5b03db74d02b832cd0155e69110494340500923a78bca550c596feed69266d92 Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.435380 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.449897 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.567203 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config\") pod \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.567283 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc\") pod \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.567419 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmwm2\" (UniqueName: \"kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2\") pod \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.567593 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config\") pod \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\" (UID: \"fcce0305-17cb-479c-b8b5-13a0fe2a7830\") " Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.567797 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stggl\" (UniqueName: \"kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl\") pod \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\" (UID: \"22d70f3e-96b4-4ea7-aad9-6d51b521731e\") " Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.568208 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config" (OuterVolumeSpecName: "config") pod "fcce0305-17cb-479c-b8b5-13a0fe2a7830" (UID: "fcce0305-17cb-479c-b8b5-13a0fe2a7830"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.568280 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config" (OuterVolumeSpecName: "config") pod "22d70f3e-96b4-4ea7-aad9-6d51b521731e" (UID: "22d70f3e-96b4-4ea7-aad9-6d51b521731e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.568319 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fcce0305-17cb-479c-b8b5-13a0fe2a7830" (UID: "fcce0305-17cb-479c-b8b5-13a0fe2a7830"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.573370 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl" (OuterVolumeSpecName: "kube-api-access-stggl") pod "22d70f3e-96b4-4ea7-aad9-6d51b521731e" (UID: "22d70f3e-96b4-4ea7-aad9-6d51b521731e"). InnerVolumeSpecName "kube-api-access-stggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.574242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2" (OuterVolumeSpecName: "kube-api-access-pmwm2") pod "fcce0305-17cb-479c-b8b5-13a0fe2a7830" (UID: "fcce0305-17cb-479c-b8b5-13a0fe2a7830"). InnerVolumeSpecName "kube-api-access-pmwm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.669550 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stggl\" (UniqueName: \"kubernetes.io/projected/22d70f3e-96b4-4ea7-aad9-6d51b521731e-kube-api-access-stggl\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.669586 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22d70f3e-96b4-4ea7-aad9-6d51b521731e-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.669596 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.669607 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmwm2\" (UniqueName: \"kubernetes.io/projected/fcce0305-17cb-479c-b8b5-13a0fe2a7830-kube-api-access-pmwm2\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.669616 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcce0305-17cb-479c-b8b5-13a0fe2a7830-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.770375 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk"] Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.786128 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:43:45 crc kubenswrapper[4888]: W1124 00:43:45.796309 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4678859e_43e2_4d0b_ae68_323ec7239ad7.slice/crio-7c3fb792900cd2feb6472fefae4de29b62fc2a5697b0f5b9476c761a0c966088 WatchSource:0}: Error finding container 7c3fb792900cd2feb6472fefae4de29b62fc2a5697b0f5b9476c761a0c966088: Status 404 returned error can't find the container with id 7c3fb792900cd2feb6472fefae4de29b62fc2a5697b0f5b9476c761a0c966088 Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.802357 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x5tvh"] Nov 24 00:43:45 crc kubenswrapper[4888]: W1124 00:43:45.806714 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod053aa355_b781_49f4_8e8a_80ae64c4de62.slice/crio-2ed279afedf0fa008cfbff77741cb121843ef3542d2f5a8851f6c260075af3fe WatchSource:0}: Error finding container 2ed279afedf0fa008cfbff77741cb121843ef3542d2f5a8851f6c260075af3fe: Status 404 returned error can't find the container with id 2ed279afedf0fa008cfbff77741cb121843ef3542d2f5a8851f6c260075af3fe Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.894747 4888 generic.go:334] "Generic (PLEG): container finished" podID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerID="90b44a9ed1bd11c2486e363d5459c1546ae5820aa10a394e962786bf859500a5" exitCode=0 Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.894890 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" event={"ID":"6df49d3e-c85e-4969-adf0-9aee1f672f4a","Type":"ContainerDied","Data":"90b44a9ed1bd11c2486e363d5459c1546ae5820aa10a394e962786bf859500a5"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.897426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fca78ec3-fec9-4c3e-aee9-75d3e049c668","Type":"ContainerStarted","Data":"14e7445212dfe4c977edebbd6c7ed27261fab7ab15011d2e61928e5b905caa21"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.902564 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" event={"ID":"22d70f3e-96b4-4ea7-aad9-6d51b521731e","Type":"ContainerDied","Data":"b689da6b06c87aac33eec626997870c7725ce92e36828a3a709f592c97b72ecd"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.902606 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fqcm7" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.910992 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerStarted","Data":"5b03db74d02b832cd0155e69110494340500923a78bca550c596feed69266d92"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.919882 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4678859e-43e2-4d0b-ae68-323ec7239ad7","Type":"ContainerStarted","Data":"7c3fb792900cd2feb6472fefae4de29b62fc2a5697b0f5b9476c761a0c966088"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.931762 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dcb56c6bb-bgkr9" event={"ID":"8307845f-a311-4346-8012-516220cf33b9","Type":"ContainerStarted","Data":"bdc71495dabdcf91eb964c814c662835c523a55bcc31901e1971d833486ff782"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.931837 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dcb56c6bb-bgkr9" event={"ID":"8307845f-a311-4346-8012-516220cf33b9","Type":"ContainerStarted","Data":"e31cf8ad4854e4095f88d5b6d1aae552b3c07afffeea18dec8772848503510e2"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.934450 4888 generic.go:334] "Generic (PLEG): container finished" podID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerID="20297c07becf40d972521b46b1c40eca97303983fecf406720a8ff1ee2ca6d21" exitCode=0 Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.934525 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" event={"ID":"79b310c2-f53b-401e-89e1-60e7cec6d274","Type":"ContainerDied","Data":"20297c07becf40d972521b46b1c40eca97303983fecf406720a8ff1ee2ca6d21"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.935734 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" event={"ID":"a05abc8e-614d-4cb7-8eda-35f0698a756c","Type":"ContainerStarted","Data":"ad862167f5b741e3cbf0aec4de03dd576e229194314b1c46841b18a4f48e176c"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.937297 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" event={"ID":"fcce0305-17cb-479c-b8b5-13a0fe2a7830","Type":"ContainerDied","Data":"fa543fe704278f1402819db54a1efdbba05681d5e4d6edef7c82783185668694"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.937428 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z7j49" Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.942098 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x5tvh" event={"ID":"053aa355-b781-49f4-8e8a-80ae64c4de62","Type":"ContainerStarted","Data":"2ed279afedf0fa008cfbff77741cb121843ef3542d2f5a8851f6c260075af3fe"} Nov 24 00:43:45 crc kubenswrapper[4888]: I1124 00:43:45.983539 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5dcb56c6bb-bgkr9" podStartSLOduration=12.98351382 podStartE2EDuration="12.98351382s" podCreationTimestamp="2025-11-24 00:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:43:45.94990098 +0000 UTC m=+1128.532585024" watchObservedRunningTime="2025-11-24 00:43:45.98351382 +0000 UTC m=+1128.566197854" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.042890 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.052951 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fqcm7"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.068173 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.075335 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z7j49"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.229982 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-g6h6d"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.279983 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22d70f3e-96b4-4ea7-aad9-6d51b521731e" path="/var/lib/kubelet/pods/22d70f3e-96b4-4ea7-aad9-6d51b521731e/volumes" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.280353 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcce0305-17cb-479c-b8b5-13a0fe2a7830" path="/var/lib/kubelet/pods/fcce0305-17cb-479c-b8b5-13a0fe2a7830/volumes" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.952865 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" event={"ID":"79b310c2-f53b-401e-89e1-60e7cec6d274","Type":"ContainerStarted","Data":"6b371b781956b24456eab8a1d3ceff2ec95a505625668610c0de123afb9ab554"} Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.953198 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.955428 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" event={"ID":"6df49d3e-c85e-4969-adf0-9aee1f672f4a","Type":"ContainerStarted","Data":"3d48bbdf118735c7220919ce837f3123047a91dd1af5dbde9639dbcbb1897d7b"} Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.955551 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.959284 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6h6d" event={"ID":"fb427d6c-c0ae-454c-8c53-71d20209ea47","Type":"ContainerStarted","Data":"70be6d59f43383040a94f2c0322fa2784f5249b9cfce0353d005ae181c1c5b61"} Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.968921 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" podStartSLOduration=2.801994755 podStartE2EDuration="20.968905678s" podCreationTimestamp="2025-11-24 00:43:26 +0000 UTC" firstStartedPulling="2025-11-24 00:43:27.013732634 +0000 UTC m=+1109.596416678" lastFinishedPulling="2025-11-24 00:43:45.180643557 +0000 UTC m=+1127.763327601" observedRunningTime="2025-11-24 00:43:46.967254952 +0000 UTC m=+1129.549938996" watchObservedRunningTime="2025-11-24 00:43:46.968905678 +0000 UTC m=+1129.551589712" Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.988560 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 00:43:46 crc kubenswrapper[4888]: I1124 00:43:46.992427 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" podStartSLOduration=6.179248575 podStartE2EDuration="20.992416436s" podCreationTimestamp="2025-11-24 00:43:26 +0000 UTC" firstStartedPulling="2025-11-24 00:43:30.086260192 +0000 UTC m=+1112.668944236" lastFinishedPulling="2025-11-24 00:43:44.899428053 +0000 UTC m=+1127.482112097" observedRunningTime="2025-11-24 00:43:46.982051446 +0000 UTC m=+1129.564735500" watchObservedRunningTime="2025-11-24 00:43:46.992416436 +0000 UTC m=+1129.575100480" Nov 24 00:43:47 crc kubenswrapper[4888]: I1124 00:43:47.203803 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 00:43:47 crc kubenswrapper[4888]: I1124 00:43:47.970522 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"728a0118-33ee-4fb3-b8e2-198e075adc31","Type":"ContainerStarted","Data":"fb7b26c0588e877a1dfecab28f833899b9d256ab85c29f19e42ab49c7e24d6ee"} Nov 24 00:43:47 crc kubenswrapper[4888]: I1124 00:43:47.973518 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cdce96a2-6d42-4761-9170-ef61211cfc92","Type":"ContainerStarted","Data":"b96b16669b550282ba67e406331c2253f42c838af022ce80d0e5c75bb8d57d38"} Nov 24 00:43:51 crc kubenswrapper[4888]: I1124 00:43:51.497333 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:51 crc kubenswrapper[4888]: I1124 00:43:51.781631 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:43:51 crc kubenswrapper[4888]: I1124 00:43:51.835983 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:52 crc kubenswrapper[4888]: I1124 00:43:52.004347 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="dnsmasq-dns" containerID="cri-o://6b371b781956b24456eab8a1d3ceff2ec95a505625668610c0de123afb9ab554" gracePeriod=10 Nov 24 00:43:53 crc kubenswrapper[4888]: I1124 00:43:53.038712 4888 generic.go:334] "Generic (PLEG): container finished" podID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerID="6b371b781956b24456eab8a1d3ceff2ec95a505625668610c0de123afb9ab554" exitCode=0 Nov 24 00:43:53 crc kubenswrapper[4888]: I1124 00:43:53.038861 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" event={"ID":"79b310c2-f53b-401e-89e1-60e7cec6d274","Type":"ContainerDied","Data":"6b371b781956b24456eab8a1d3ceff2ec95a505625668610c0de123afb9ab554"} Nov 24 00:43:54 crc kubenswrapper[4888]: I1124 00:43:54.051954 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:54 crc kubenswrapper[4888]: I1124 00:43:54.052061 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:54 crc kubenswrapper[4888]: I1124 00:43:54.066127 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:55 crc kubenswrapper[4888]: I1124 00:43:55.063186 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5dcb56c6bb-bgkr9" Nov 24 00:43:55 crc kubenswrapper[4888]: I1124 00:43:55.139652 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:43:56 crc kubenswrapper[4888]: I1124 00:43:56.497066 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: connect: connection refused" Nov 24 00:43:56 crc kubenswrapper[4888]: I1124 00:43:56.964989 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.079030 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" event={"ID":"79b310c2-f53b-401e-89e1-60e7cec6d274","Type":"ContainerDied","Data":"8847fef06b095a423bbbabc76e85bc7430348ae39f060f87cfac3a0ce4e08dbf"} Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.079076 4888 scope.go:117] "RemoveContainer" containerID="6b371b781956b24456eab8a1d3ceff2ec95a505625668610c0de123afb9ab554" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.079224 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fzvnt" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.094360 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc\") pod \"79b310c2-f53b-401e-89e1-60e7cec6d274\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.094461 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config\") pod \"79b310c2-f53b-401e-89e1-60e7cec6d274\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.094507 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssh28\" (UniqueName: \"kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28\") pod \"79b310c2-f53b-401e-89e1-60e7cec6d274\" (UID: \"79b310c2-f53b-401e-89e1-60e7cec6d274\") " Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.098121 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28" (OuterVolumeSpecName: "kube-api-access-ssh28") pod "79b310c2-f53b-401e-89e1-60e7cec6d274" (UID: "79b310c2-f53b-401e-89e1-60e7cec6d274"). InnerVolumeSpecName "kube-api-access-ssh28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.133211 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config" (OuterVolumeSpecName: "config") pod "79b310c2-f53b-401e-89e1-60e7cec6d274" (UID: "79b310c2-f53b-401e-89e1-60e7cec6d274"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.138045 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79b310c2-f53b-401e-89e1-60e7cec6d274" (UID: "79b310c2-f53b-401e-89e1-60e7cec6d274"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.196987 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.197041 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b310c2-f53b-401e-89e1-60e7cec6d274-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.197051 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssh28\" (UniqueName: \"kubernetes.io/projected/79b310c2-f53b-401e-89e1-60e7cec6d274-kube-api-access-ssh28\") on node \"crc\" DevicePath \"\"" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.312236 4888 scope.go:117] "RemoveContainer" containerID="20297c07becf40d972521b46b1c40eca97303983fecf406720a8ff1ee2ca6d21" Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.430914 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:57 crc kubenswrapper[4888]: I1124 00:43:57.444041 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fzvnt"] Nov 24 00:43:58 crc kubenswrapper[4888]: I1124 00:43:58.089829 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fca78ec3-fec9-4c3e-aee9-75d3e049c668","Type":"ContainerStarted","Data":"01ef2be8e69ee0f18c912829c9ee63ae295ed652dc2592788775307090e08fda"} Nov 24 00:43:58 crc kubenswrapper[4888]: I1124 00:43:58.090150 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 00:43:58 crc kubenswrapper[4888]: I1124 00:43:58.118843 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.881614187 podStartE2EDuration="28.11882109s" podCreationTimestamp="2025-11-24 00:43:30 +0000 UTC" firstStartedPulling="2025-11-24 00:43:45.077243656 +0000 UTC m=+1127.659927700" lastFinishedPulling="2025-11-24 00:43:56.314450559 +0000 UTC m=+1138.897134603" observedRunningTime="2025-11-24 00:43:58.117481213 +0000 UTC m=+1140.700165257" watchObservedRunningTime="2025-11-24 00:43:58.11882109 +0000 UTC m=+1140.701505134" Nov 24 00:43:58 crc kubenswrapper[4888]: I1124 00:43:58.258112 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" path="/var/lib/kubelet/pods/79b310c2-f53b-401e-89e1-60e7cec6d274/volumes" Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.100249 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" event={"ID":"a05abc8e-614d-4cb7-8eda-35f0698a756c","Type":"ContainerStarted","Data":"29e5829d7c79cc2d4c43ddd0dfe282792603971a5e2e83e36b18a4c53b15cb70"} Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.102404 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7","Type":"ContainerStarted","Data":"c0059efbf82a3768634cff02c16ce5f8691a0f668c5795a9e7092060d720b6d4"} Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.105568 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1c29616-6ecc-4f98-9849-b04242cf0a66","Type":"ContainerStarted","Data":"33e4e73d348542615d4bbb6f251dc131ede7d7f1c3e8615b680f633d49a43d0f"} Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.107287 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cdce96a2-6d42-4761-9170-ef61211cfc92","Type":"ContainerStarted","Data":"7d1f415e4425bbe02242ddad289172653bdd51dfabbc3b226952bfc1beda00c8"} Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.109926 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"728a0118-33ee-4fb3-b8e2-198e075adc31","Type":"ContainerStarted","Data":"a6e87ae6597ae9529e80c5e7541c1caa9b9169a279c1db0707e59dd90055ef3e"} Nov 24 00:43:59 crc kubenswrapper[4888]: I1124 00:43:59.129188 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-clzxk" podStartSLOduration=14.972918988 podStartE2EDuration="26.129160176s" podCreationTimestamp="2025-11-24 00:43:33 +0000 UTC" firstStartedPulling="2025-11-24 00:43:45.786891142 +0000 UTC m=+1128.369575186" lastFinishedPulling="2025-11-24 00:43:56.94313233 +0000 UTC m=+1139.525816374" observedRunningTime="2025-11-24 00:43:59.124551197 +0000 UTC m=+1141.707235281" watchObservedRunningTime="2025-11-24 00:43:59.129160176 +0000 UTC m=+1141.711844250" Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.121181 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x5tvh" event={"ID":"053aa355-b781-49f4-8e8a-80ae64c4de62","Type":"ContainerStarted","Data":"9041943ad3accee1e391f7a4323d7d02db1b44d68b537b64490277bff47a7065"} Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.121517 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-x5tvh" Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.123723 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb427d6c-c0ae-454c-8c53-71d20209ea47" containerID="bfcff5498465b0dd3e7cb375e4da110970efd858096800c0136df60d9bf43fc8" exitCode=0 Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.123863 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6h6d" event={"ID":"fb427d6c-c0ae-454c-8c53-71d20209ea47","Type":"ContainerDied","Data":"bfcff5498465b0dd3e7cb375e4da110970efd858096800c0136df60d9bf43fc8"} Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.126289 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4678859e-43e2-4d0b-ae68-323ec7239ad7","Type":"ContainerStarted","Data":"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988"} Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.126419 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.128104 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerStarted","Data":"77f92ef1d28bc7ff0790ff6bf0779443bddf3af848718975cf6a1958e090c6bb"} Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.130542 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerStarted","Data":"d49093346fd5a6787c95e5e78049f40011ec54168f4dcdf7c2296c2c4d02c337"} Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.145200 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-x5tvh" podStartSLOduration=12.00850571 podStartE2EDuration="23.145182471s" podCreationTimestamp="2025-11-24 00:43:37 +0000 UTC" firstStartedPulling="2025-11-24 00:43:45.809858285 +0000 UTC m=+1128.392542329" lastFinishedPulling="2025-11-24 00:43:56.946535056 +0000 UTC m=+1139.529219090" observedRunningTime="2025-11-24 00:44:00.135705766 +0000 UTC m=+1142.718389800" watchObservedRunningTime="2025-11-24 00:44:00.145182471 +0000 UTC m=+1142.727866505" Nov 24 00:44:00 crc kubenswrapper[4888]: I1124 00:44:00.204565 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.02919338 podStartE2EDuration="28.20454849s" podCreationTimestamp="2025-11-24 00:43:32 +0000 UTC" firstStartedPulling="2025-11-24 00:43:45.798839796 +0000 UTC m=+1128.381523840" lastFinishedPulling="2025-11-24 00:43:57.974194906 +0000 UTC m=+1140.556878950" observedRunningTime="2025-11-24 00:44:00.197519254 +0000 UTC m=+1142.780203318" watchObservedRunningTime="2025-11-24 00:44:00.20454849 +0000 UTC m=+1142.787232534" Nov 24 00:44:02 crc kubenswrapper[4888]: I1124 00:44:02.152179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerStarted","Data":"79592bc08a061c8ba660552008cbf8fb97997b70e608e100de19683b1c0dc9a8"} Nov 24 00:44:02 crc kubenswrapper[4888]: I1124 00:44:02.157208 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6h6d" event={"ID":"fb427d6c-c0ae-454c-8c53-71d20209ea47","Type":"ContainerStarted","Data":"d6baae22da13f536d198d0a8b0c89b7d379e63bfc28cb6d6c0d4d195982b8389"} Nov 24 00:44:03 crc kubenswrapper[4888]: I1124 00:44:03.174151 4888 generic.go:334] "Generic (PLEG): container finished" podID="d1c29616-6ecc-4f98-9849-b04242cf0a66" containerID="33e4e73d348542615d4bbb6f251dc131ede7d7f1c3e8615b680f633d49a43d0f" exitCode=0 Nov 24 00:44:03 crc kubenswrapper[4888]: I1124 00:44:03.174283 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1c29616-6ecc-4f98-9849-b04242cf0a66","Type":"ContainerDied","Data":"33e4e73d348542615d4bbb6f251dc131ede7d7f1c3e8615b680f633d49a43d0f"} Nov 24 00:44:03 crc kubenswrapper[4888]: I1124 00:44:03.188005 4888 generic.go:334] "Generic (PLEG): container finished" podID="9a963d44-99cd-45d4-afd8-9c9c7bfb95e7" containerID="c0059efbf82a3768634cff02c16ce5f8691a0f668c5795a9e7092060d720b6d4" exitCode=0 Nov 24 00:44:03 crc kubenswrapper[4888]: I1124 00:44:03.188337 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7","Type":"ContainerDied","Data":"c0059efbf82a3768634cff02c16ce5f8691a0f668c5795a9e7092060d720b6d4"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.198841 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a963d44-99cd-45d4-afd8-9c9c7bfb95e7","Type":"ContainerStarted","Data":"791e1d23874e0bb4f542575aa425348ad0926208583c36a3b181a9db8b7fd5aa"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.201620 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1c29616-6ecc-4f98-9849-b04242cf0a66","Type":"ContainerStarted","Data":"9d365df170a2770bfee9f5b395fda05b67a0b533d9049a29f494c26207bedfc4"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.206266 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6h6d" event={"ID":"fb427d6c-c0ae-454c-8c53-71d20209ea47","Type":"ContainerStarted","Data":"34e7e1fdbed2ac365cf406af42c67f7117dc62282c68fb0c22694cbe11970b1f"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.206914 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.207050 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.210376 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cdce96a2-6d42-4761-9170-ef61211cfc92","Type":"ContainerStarted","Data":"2d59ed001957af838152ab8c6ca87f20e450d74f872fc00ce869edf2b6001ca7"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.214372 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"728a0118-33ee-4fb3-b8e2-198e075adc31","Type":"ContainerStarted","Data":"d13f11213225544106409bda618d7bff87de980469b6472630d6af2e2d002d0b"} Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.236674 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.844902022 podStartE2EDuration="35.236652494s" podCreationTimestamp="2025-11-24 00:43:29 +0000 UTC" firstStartedPulling="2025-11-24 00:43:44.551249045 +0000 UTC m=+1127.133933129" lastFinishedPulling="2025-11-24 00:43:56.942999537 +0000 UTC m=+1139.525683601" observedRunningTime="2025-11-24 00:44:04.230162402 +0000 UTC m=+1146.812846466" watchObservedRunningTime="2025-11-24 00:44:04.236652494 +0000 UTC m=+1146.819336548" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.277300 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-g6h6d" podStartSLOduration=16.81074119 podStartE2EDuration="27.27727707s" podCreationTimestamp="2025-11-24 00:43:37 +0000 UTC" firstStartedPulling="2025-11-24 00:43:46.242782701 +0000 UTC m=+1128.825466745" lastFinishedPulling="2025-11-24 00:43:56.709318571 +0000 UTC m=+1139.292002625" observedRunningTime="2025-11-24 00:44:04.266728805 +0000 UTC m=+1146.849412859" watchObservedRunningTime="2025-11-24 00:44:04.27727707 +0000 UTC m=+1146.859961124" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.327370 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.935736663 podStartE2EDuration="27.32734661s" podCreationTimestamp="2025-11-24 00:43:37 +0000 UTC" firstStartedPulling="2025-11-24 00:43:47.707907906 +0000 UTC m=+1130.290591960" lastFinishedPulling="2025-11-24 00:44:03.099517873 +0000 UTC m=+1145.682201907" observedRunningTime="2025-11-24 00:44:04.301992971 +0000 UTC m=+1146.884677025" watchObservedRunningTime="2025-11-24 00:44:04.32734661 +0000 UTC m=+1146.910030664" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.354653 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.976799731 podStartE2EDuration="25.354626913s" podCreationTimestamp="2025-11-24 00:43:39 +0000 UTC" firstStartedPulling="2025-11-24 00:43:47.707993328 +0000 UTC m=+1130.290677402" lastFinishedPulling="2025-11-24 00:44:03.08582054 +0000 UTC m=+1145.668504584" observedRunningTime="2025-11-24 00:44:04.328901574 +0000 UTC m=+1146.911585658" watchObservedRunningTime="2025-11-24 00:44:04.354626913 +0000 UTC m=+1146.937310957" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.361136 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.78611227 podStartE2EDuration="36.361120445s" podCreationTimestamp="2025-11-24 00:43:28 +0000 UTC" firstStartedPulling="2025-11-24 00:43:35.134609525 +0000 UTC m=+1117.717293579" lastFinishedPulling="2025-11-24 00:43:56.70961771 +0000 UTC m=+1139.292301754" observedRunningTime="2025-11-24 00:44:04.350205429 +0000 UTC m=+1146.932889493" watchObservedRunningTime="2025-11-24 00:44:04.361120445 +0000 UTC m=+1146.943804489" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.496874 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 00:44:04 crc kubenswrapper[4888]: I1124 00:44:04.562592 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.227801 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.276344 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.609266 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:05 crc kubenswrapper[4888]: E1124 00:44:05.609584 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="dnsmasq-dns" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.609599 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="dnsmasq-dns" Nov 24 00:44:05 crc kubenswrapper[4888]: E1124 00:44:05.609612 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="init" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.609618 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="init" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.609773 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b310c2-f53b-401e-89e1-60e7cec6d274" containerName="dnsmasq-dns" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.610676 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.612591 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.622950 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.646617 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.653553 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4s282"] Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.654920 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.657052 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.660085 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4s282"] Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.703054 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.768802 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctp4l\" (UniqueName: \"kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.768869 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovn-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.768910 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovs-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.768943 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-combined-ca-bundle\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769098 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769185 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmn2k\" (UniqueName: \"kubernetes.io/projected/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-kube-api-access-kmn2k\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769285 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-config\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769313 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769488 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.769571 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.878944 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879025 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879093 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctp4l\" (UniqueName: \"kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879134 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovn-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879195 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovs-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879255 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-combined-ca-bundle\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879307 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879350 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmn2k\" (UniqueName: \"kubernetes.io/projected/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-kube-api-access-kmn2k\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879417 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-config\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879484 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovn-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879799 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-ovs-rundir\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.879999 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.880282 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-config\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.880356 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.880552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.888982 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-combined-ca-bundle\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.894436 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.899451 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctp4l\" (UniqueName: \"kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l\") pod \"dnsmasq-dns-7f896c8c65-m97l7\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.902340 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmn2k\" (UniqueName: \"kubernetes.io/projected/6f28f199-4a0d-4c00-8aef-c5219f6fcc10-kube-api-access-kmn2k\") pod \"ovn-controller-metrics-4s282\" (UID: \"6f28f199-4a0d-4c00-8aef-c5219f6fcc10\") " pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.933418 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.973145 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:05 crc kubenswrapper[4888]: I1124 00:44:05.977776 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4s282" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.000307 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.002116 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.003977 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.004141 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.084722 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.084893 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.084970 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.085145 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.085242 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw96d\" (UniqueName: \"kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.186827 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.186910 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.186946 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.187031 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.187079 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw96d\" (UniqueName: \"kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.188173 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.188676 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.189137 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.189241 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.205605 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw96d\" (UniqueName: \"kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d\") pod \"dnsmasq-dns-86db49b7ff-hbcnd\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.226950 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.239541 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.346065 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.398468 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.482020 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:06 crc kubenswrapper[4888]: W1124 00:44:06.486909 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod836bf14a_da67_4acd_ab41_e26a7ffcd1c2.slice/crio-9e4cf5f18212a8d322559e11fc61d802a6e8069ef6d65dd3820380da110651da WatchSource:0}: Error finding container 9e4cf5f18212a8d322559e11fc61d802a6e8069ef6d65dd3820380da110651da: Status 404 returned error can't find the container with id 9e4cf5f18212a8d322559e11fc61d802a6e8069ef6d65dd3820380da110651da Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.544523 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4s282"] Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.626870 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.639272 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.642058 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.645789 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.646091 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.646211 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.646318 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-ldgtp" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805207 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805313 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-scripts\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805340 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-config\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805359 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805380 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805403 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.805418 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49qsz\" (UniqueName: \"kubernetes.io/projected/2c37246a-89a8-4f28-95f9-149b6c9f1790-kube-api-access-49qsz\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.845160 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:06 crc kubenswrapper[4888]: W1124 00:44:06.846587 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40c94c41_2c17_4cb0_8f85_da8fa2ff942b.slice/crio-63b7d95f886737acfa8eab0a45584d8ee7ef01d6e0383dec0fdf462d93c446dc WatchSource:0}: Error finding container 63b7d95f886737acfa8eab0a45584d8ee7ef01d6e0383dec0fdf462d93c446dc: Status 404 returned error can't find the container with id 63b7d95f886737acfa8eab0a45584d8ee7ef01d6e0383dec0fdf462d93c446dc Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908466 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-scripts\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908772 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-config\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908796 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908829 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908851 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908866 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49qsz\" (UniqueName: \"kubernetes.io/projected/2c37246a-89a8-4f28-95f9-149b6c9f1790-kube-api-access-49qsz\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.908945 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.913689 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.914472 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-scripts\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.917213 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.917454 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c37246a-89a8-4f28-95f9-149b6c9f1790-config\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.928520 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.935558 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c37246a-89a8-4f28-95f9-149b6c9f1790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.939208 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49qsz\" (UniqueName: \"kubernetes.io/projected/2c37246a-89a8-4f28-95f9-149b6c9f1790-kube-api-access-49qsz\") pod \"ovn-northd-0\" (UID: \"2c37246a-89a8-4f28-95f9-149b6c9f1790\") " pod="openstack/ovn-northd-0" Nov 24 00:44:06 crc kubenswrapper[4888]: I1124 00:44:06.983109 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.254500 4888 generic.go:334] "Generic (PLEG): container finished" podID="836bf14a-da67-4acd-ab41-e26a7ffcd1c2" containerID="3775b7f5ee0469f3a29179f0b215fbabc1c1358be6c5ae3617fbc97aa9be7be5" exitCode=0 Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.254577 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" event={"ID":"836bf14a-da67-4acd-ab41-e26a7ffcd1c2","Type":"ContainerDied","Data":"3775b7f5ee0469f3a29179f0b215fbabc1c1358be6c5ae3617fbc97aa9be7be5"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.254918 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" event={"ID":"836bf14a-da67-4acd-ab41-e26a7ffcd1c2","Type":"ContainerStarted","Data":"9e4cf5f18212a8d322559e11fc61d802a6e8069ef6d65dd3820380da110651da"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.257358 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4s282" event={"ID":"6f28f199-4a0d-4c00-8aef-c5219f6fcc10","Type":"ContainerStarted","Data":"89249e1446ec729a76dd4f261a43e63cf7498d2880c3286301e89232542888cc"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.257394 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4s282" event={"ID":"6f28f199-4a0d-4c00-8aef-c5219f6fcc10","Type":"ContainerStarted","Data":"58804cbd612874371b26440ad8f868c92d683a21883b25089b872ba358e7f88e"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.259879 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerStarted","Data":"4659b25c657f59d8328b3cfca7d066ab83a2dcd32f12a5e7d6db5473852fc206"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.259924 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerStarted","Data":"63b7d95f886737acfa8eab0a45584d8ee7ef01d6e0383dec0fdf462d93c446dc"} Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.300072 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4s282" podStartSLOduration=2.300050987 podStartE2EDuration="2.300050987s" podCreationTimestamp="2025-11-24 00:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:44:07.284146202 +0000 UTC m=+1149.866830256" watchObservedRunningTime="2025-11-24 00:44:07.300050987 +0000 UTC m=+1149.882735031" Nov 24 00:44:07 crc kubenswrapper[4888]: E1124 00:44:07.456217 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40c94c41_2c17_4cb0_8f85_da8fa2ff942b.slice/crio-conmon-4659b25c657f59d8328b3cfca7d066ab83a2dcd32f12a5e7d6db5473852fc206.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.554803 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.699142 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.847639 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctp4l\" (UniqueName: \"kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l\") pod \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.847778 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config\") pod \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.847821 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc\") pod \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.847884 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb\") pod \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\" (UID: \"836bf14a-da67-4acd-ab41-e26a7ffcd1c2\") " Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.875336 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "836bf14a-da67-4acd-ab41-e26a7ffcd1c2" (UID: "836bf14a-da67-4acd-ab41-e26a7ffcd1c2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.875690 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "836bf14a-da67-4acd-ab41-e26a7ffcd1c2" (UID: "836bf14a-da67-4acd-ab41-e26a7ffcd1c2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.878532 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config" (OuterVolumeSpecName: "config") pod "836bf14a-da67-4acd-ab41-e26a7ffcd1c2" (UID: "836bf14a-da67-4acd-ab41-e26a7ffcd1c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.949484 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.949518 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.949529 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:07 crc kubenswrapper[4888]: I1124 00:44:07.955545 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l" (OuterVolumeSpecName: "kube-api-access-ctp4l") pod "836bf14a-da67-4acd-ab41-e26a7ffcd1c2" (UID: "836bf14a-da67-4acd-ab41-e26a7ffcd1c2"). InnerVolumeSpecName "kube-api-access-ctp4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.052007 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctp4l\" (UniqueName: \"kubernetes.io/projected/836bf14a-da67-4acd-ab41-e26a7ffcd1c2-kube-api-access-ctp4l\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.279070 4888 generic.go:334] "Generic (PLEG): container finished" podID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerID="79592bc08a061c8ba660552008cbf8fb97997b70e608e100de19683b1c0dc9a8" exitCode=0 Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.279171 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerDied","Data":"79592bc08a061c8ba660552008cbf8fb97997b70e608e100de19683b1c0dc9a8"} Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.284538 4888 generic.go:334] "Generic (PLEG): container finished" podID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerID="4659b25c657f59d8328b3cfca7d066ab83a2dcd32f12a5e7d6db5473852fc206" exitCode=0 Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.284682 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerDied","Data":"4659b25c657f59d8328b3cfca7d066ab83a2dcd32f12a5e7d6db5473852fc206"} Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.286387 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c37246a-89a8-4f28-95f9-149b6c9f1790","Type":"ContainerStarted","Data":"fcffdd0fac5ca9f1822de8d5eacd2bf52a0723a0877b97469c9bcdf0569dc894"} Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.289584 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" event={"ID":"836bf14a-da67-4acd-ab41-e26a7ffcd1c2","Type":"ContainerDied","Data":"9e4cf5f18212a8d322559e11fc61d802a6e8069ef6d65dd3820380da110651da"} Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.289636 4888 scope.go:117] "RemoveContainer" containerID="3775b7f5ee0469f3a29179f0b215fbabc1c1358be6c5ae3617fbc97aa9be7be5" Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.289936 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-m97l7" Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.528505 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:08 crc kubenswrapper[4888]: I1124 00:44:08.537306 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-m97l7"] Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.301424 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerStarted","Data":"a0a02c64906e577a1c39a733e949ec7939c76cec1f1699748be4e19893560327"} Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.302722 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.303938 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c37246a-89a8-4f28-95f9-149b6c9f1790","Type":"ContainerStarted","Data":"82945b20633d3c9392cfee0ef2e08769c3d5da0be9431eff664aa76c0feb3807"} Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.304025 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c37246a-89a8-4f28-95f9-149b6c9f1790","Type":"ContainerStarted","Data":"c8e6950e3315b4ab8a1fe2bd64036c754fa872a3a0b2a85183fc05819813371c"} Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.304852 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.329638 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" podStartSLOduration=4.329619886 podStartE2EDuration="4.329619886s" podCreationTimestamp="2025-11-24 00:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:44:09.327365583 +0000 UTC m=+1151.910049627" watchObservedRunningTime="2025-11-24 00:44:09.329619886 +0000 UTC m=+1151.912303930" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.350632 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.108393972 podStartE2EDuration="3.350614423s" podCreationTimestamp="2025-11-24 00:44:06 +0000 UTC" firstStartedPulling="2025-11-24 00:44:07.59662643 +0000 UTC m=+1150.179310474" lastFinishedPulling="2025-11-24 00:44:08.838846891 +0000 UTC m=+1151.421530925" observedRunningTime="2025-11-24 00:44:09.347057404 +0000 UTC m=+1151.929741468" watchObservedRunningTime="2025-11-24 00:44:09.350614423 +0000 UTC m=+1151.933298467" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.374042 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.374211 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 00:44:09 crc kubenswrapper[4888]: I1124 00:44:09.494444 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.258746 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="836bf14a-da67-4acd-ab41-e26a7ffcd1c2" path="/var/lib/kubelet/pods/836bf14a-da67-4acd-ab41-e26a7ffcd1c2/volumes" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.393497 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.762689 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-aa04-account-create-c6q7c"] Nov 24 00:44:10 crc kubenswrapper[4888]: E1124 00:44:10.763288 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836bf14a-da67-4acd-ab41-e26a7ffcd1c2" containerName="init" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.763312 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="836bf14a-da67-4acd-ab41-e26a7ffcd1c2" containerName="init" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.763652 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="836bf14a-da67-4acd-ab41-e26a7ffcd1c2" containerName="init" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.764706 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.766936 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.772469 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-aa04-account-create-c6q7c"] Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.807934 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktcmp\" (UniqueName: \"kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.808225 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.808920 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jpqvm"] Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.810135 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.818667 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jpqvm"] Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.884371 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.884417 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.910838 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.911040 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.911147 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zddl2\" (UniqueName: \"kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.911201 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktcmp\" (UniqueName: \"kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.911668 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.937689 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktcmp\" (UniqueName: \"kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp\") pod \"keystone-aa04-account-create-c6q7c\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:10 crc kubenswrapper[4888]: I1124 00:44:10.969271 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.012085 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zddl2\" (UniqueName: \"kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.012249 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.013775 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.028895 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zddl2\" (UniqueName: \"kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2\") pod \"keystone-db-create-jpqvm\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.099214 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.100858 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wx4qp"] Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.102865 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.114401 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.114465 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc6jf\" (UniqueName: \"kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.120225 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wx4qp"] Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.129481 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.200579 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3726-account-create-7nrm5"] Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.201658 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.205469 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.226562 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.230976 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3726-account-create-7nrm5"] Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.244856 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hrc2\" (UniqueName: \"kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.244981 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.245025 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc6jf\" (UniqueName: \"kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.246294 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.272591 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc6jf\" (UniqueName: \"kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf\") pod \"placement-db-create-wx4qp\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.347069 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.347566 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hrc2\" (UniqueName: \"kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.349211 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.366463 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hrc2\" (UniqueName: \"kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2\") pod \"placement-3726-account-create-7nrm5\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.438343 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.565575 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.586610 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.592719 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-aa04-account-create-c6q7c"] Nov 24 00:44:11 crc kubenswrapper[4888]: W1124 00:44:11.607750 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1af2ab20_ea96_42e5_9aa1_3a319b4a5630.slice/crio-db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31 WatchSource:0}: Error finding container db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31: Status 404 returned error can't find the container with id db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31 Nov 24 00:44:11 crc kubenswrapper[4888]: I1124 00:44:11.735986 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jpqvm"] Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.037706 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wx4qp"] Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.104435 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3726-account-create-7nrm5"] Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.350516 4888 generic.go:334] "Generic (PLEG): container finished" podID="1af2ab20-ea96-42e5-9aa1-3a319b4a5630" containerID="89c2bc8e71de1d786fdf59a7f7525d85409394bcf748b023ab534a3695febbb0" exitCode=0 Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.350555 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aa04-account-create-c6q7c" event={"ID":"1af2ab20-ea96-42e5-9aa1-3a319b4a5630","Type":"ContainerDied","Data":"89c2bc8e71de1d786fdf59a7f7525d85409394bcf748b023ab534a3695febbb0"} Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.350901 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aa04-account-create-c6q7c" event={"ID":"1af2ab20-ea96-42e5-9aa1-3a319b4a5630","Type":"ContainerStarted","Data":"db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31"} Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.352757 4888 generic.go:334] "Generic (PLEG): container finished" podID="47fc5340-6cad-4102-8d70-c57ea387f607" containerID="658b6f8d0862b4c0e48c2fc7ce2229cc5c1eeab90483ccacbc617f680dd43442" exitCode=0 Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.352829 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jpqvm" event={"ID":"47fc5340-6cad-4102-8d70-c57ea387f607","Type":"ContainerDied","Data":"658b6f8d0862b4c0e48c2fc7ce2229cc5c1eeab90483ccacbc617f680dd43442"} Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.352869 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jpqvm" event={"ID":"47fc5340-6cad-4102-8d70-c57ea387f607","Type":"ContainerStarted","Data":"d15e13697dcb613e4756b9cda9e2a14d47be37499c5bf605c5cf1c7c1e10539a"} Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.854750 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vm5vz"] Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.856384 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.861705 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vm5vz"] Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.893760 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.893858 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhdcb\" (UniqueName: \"kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.995679 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.995780 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhdcb\" (UniqueName: \"kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:12 crc kubenswrapper[4888]: I1124 00:44:12.996683 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.015317 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0298-account-create-bt4k8"] Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.016664 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.024161 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.050284 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0298-account-create-bt4k8"] Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.058307 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhdcb\" (UniqueName: \"kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb\") pod \"mysqld-exporter-openstack-db-create-vm5vz\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.087050 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.104469 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fp28\" (UniqueName: \"kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.104832 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.128539 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.128741 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="dnsmasq-dns" containerID="cri-o://a0a02c64906e577a1c39a733e949ec7939c76cec1f1699748be4e19893560327" gracePeriod=10 Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.132023 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.164110 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.165898 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.183716 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.190294 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.210724 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.210785 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.210894 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.210964 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fp28\" (UniqueName: \"kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.210997 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.211020 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.211036 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgphw\" (UniqueName: \"kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.211986 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.229616 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fp28\" (UniqueName: \"kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28\") pod \"mysqld-exporter-0298-account-create-bt4k8\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.313188 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.313244 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgphw\" (UniqueName: \"kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.313310 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.313338 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.313429 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.315601 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.320948 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.322436 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.324089 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.337803 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgphw\" (UniqueName: \"kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw\") pod \"dnsmasq-dns-698758b865-m5bkb\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.341755 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.374719 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerDied","Data":"a0a02c64906e577a1c39a733e949ec7939c76cec1f1699748be4e19893560327"} Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.374575 4888 generic.go:334] "Generic (PLEG): container finished" podID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerID="a0a02c64906e577a1c39a733e949ec7939c76cec1f1699748be4e19893560327" exitCode=0 Nov 24 00:44:13 crc kubenswrapper[4888]: I1124 00:44:13.489423 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.233975 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.244608 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.249529 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.250485 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.251131 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.251360 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-p75pl" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.280845 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.359761 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.359899 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-lock\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.360026 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.360196 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txkx5\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-kube-api-access-txkx5\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.360497 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-cache\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.463740 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-cache\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.463806 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.463879 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-lock\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.463936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.464005 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txkx5\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-kube-api-access-txkx5\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.464043 4888 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.464070 4888 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.464138 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift podName:9f931286-c309-4f90-8c3e-7f2b3682e68c nodeName:}" failed. No retries permitted until 2025-11-24 00:44:14.96411514 +0000 UTC m=+1157.546799184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift") pod "swift-storage-0" (UID: "9f931286-c309-4f90-8c3e-7f2b3682e68c") : configmap "swift-ring-files" not found Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.464298 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-cache\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.464419 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9f931286-c309-4f90-8c3e-7f2b3682e68c-lock\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.464488 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.489045 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txkx5\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-kube-api-access-txkx5\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.520178 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.714360 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bq8p7"] Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.715486 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.717462 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.717496 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.717678 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.730683 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bq8p7"] Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769080 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769159 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769197 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769290 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4bsl\" (UniqueName: \"kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769335 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769607 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.769658 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872046 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4bsl\" (UniqueName: \"kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872177 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872343 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872389 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872461 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872511 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.872569 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.873400 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.874170 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.874447 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.877529 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.877764 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.878732 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.889213 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4bsl\" (UniqueName: \"kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl\") pod \"swift-ring-rebalance-bq8p7\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:14 crc kubenswrapper[4888]: I1124 00:44:14.975330 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.975525 4888 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.975566 4888 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 00:44:14 crc kubenswrapper[4888]: E1124 00:44:14.975632 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift podName:9f931286-c309-4f90-8c3e-7f2b3682e68c nodeName:}" failed. No retries permitted until 2025-11-24 00:44:15.975615084 +0000 UTC m=+1158.558299118 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift") pod "swift-storage-0" (UID: "9f931286-c309-4f90-8c3e-7f2b3682e68c") : configmap "swift-ring-files" not found Nov 24 00:44:15 crc kubenswrapper[4888]: I1124 00:44:15.037766 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:15 crc kubenswrapper[4888]: I1124 00:44:15.994580 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:15 crc kubenswrapper[4888]: E1124 00:44:15.994904 4888 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 00:44:15 crc kubenswrapper[4888]: E1124 00:44:15.994957 4888 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 00:44:15 crc kubenswrapper[4888]: E1124 00:44:15.995061 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift podName:9f931286-c309-4f90-8c3e-7f2b3682e68c nodeName:}" failed. No retries permitted until 2025-11-24 00:44:17.995033044 +0000 UTC m=+1160.577717128 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift") pod "swift-storage-0" (UID: "9f931286-c309-4f90-8c3e-7f2b3682e68c") : configmap "swift-ring-files" not found Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.379479 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-7vxql"] Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.381868 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.404327 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.136:5353: connect: connection refused" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.412734 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7vxql"] Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.514702 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpb6l\" (UniqueName: \"kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.514774 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.526009 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-be14-account-create-jqd96"] Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.527704 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.531696 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.551464 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be14-account-create-jqd96"] Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.616871 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4svl\" (UniqueName: \"kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.616923 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpb6l\" (UniqueName: \"kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.616963 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.617042 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.617996 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.625605 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.656425 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.656958 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpb6l\" (UniqueName: \"kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l\") pod \"glance-db-create-7vxql\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.720878 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7vxql" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.729257 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zddl2\" (UniqueName: \"kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2\") pod \"47fc5340-6cad-4102-8d70-c57ea387f607\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.729367 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts\") pod \"47fc5340-6cad-4102-8d70-c57ea387f607\" (UID: \"47fc5340-6cad-4102-8d70-c57ea387f607\") " Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.729466 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktcmp\" (UniqueName: \"kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp\") pod \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.729586 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts\") pod \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\" (UID: \"1af2ab20-ea96-42e5-9aa1-3a319b4a5630\") " Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.730655 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47fc5340-6cad-4102-8d70-c57ea387f607" (UID: "47fc5340-6cad-4102-8d70-c57ea387f607"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.730701 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1af2ab20-ea96-42e5-9aa1-3a319b4a5630" (UID: "1af2ab20-ea96-42e5-9aa1-3a319b4a5630"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.731175 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4svl\" (UniqueName: \"kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.731324 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.731483 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47fc5340-6cad-4102-8d70-c57ea387f607-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.731497 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.734182 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.735615 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp" (OuterVolumeSpecName: "kube-api-access-ktcmp") pod "1af2ab20-ea96-42e5-9aa1-3a319b4a5630" (UID: "1af2ab20-ea96-42e5-9aa1-3a319b4a5630"). InnerVolumeSpecName "kube-api-access-ktcmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.735714 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2" (OuterVolumeSpecName: "kube-api-access-zddl2") pod "47fc5340-6cad-4102-8d70-c57ea387f607" (UID: "47fc5340-6cad-4102-8d70-c57ea387f607"). InnerVolumeSpecName "kube-api-access-zddl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.748600 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4svl\" (UniqueName: \"kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl\") pod \"glance-be14-account-create-jqd96\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.772522 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.832867 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zddl2\" (UniqueName: \"kubernetes.io/projected/47fc5340-6cad-4102-8d70-c57ea387f607-kube-api-access-zddl2\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.832906 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktcmp\" (UniqueName: \"kubernetes.io/projected/1af2ab20-ea96-42e5-9aa1-3a319b4a5630-kube-api-access-ktcmp\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:16 crc kubenswrapper[4888]: I1124 00:44:16.989007 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.143487 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw96d\" (UniqueName: \"kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d\") pod \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.143867 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb\") pod \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.143989 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config\") pod \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.144041 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc\") pod \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.144093 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb\") pod \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\" (UID: \"40c94c41-2c17-4cb0-8f85-da8fa2ff942b\") " Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.149711 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d" (OuterVolumeSpecName: "kube-api-access-qw96d") pod "40c94c41-2c17-4cb0-8f85-da8fa2ff942b" (UID: "40c94c41-2c17-4cb0-8f85-da8fa2ff942b"). InnerVolumeSpecName "kube-api-access-qw96d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.210884 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "40c94c41-2c17-4cb0-8f85-da8fa2ff942b" (UID: "40c94c41-2c17-4cb0-8f85-da8fa2ff942b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.216689 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config" (OuterVolumeSpecName: "config") pod "40c94c41-2c17-4cb0-8f85-da8fa2ff942b" (UID: "40c94c41-2c17-4cb0-8f85-da8fa2ff942b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.216996 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40c94c41-2c17-4cb0-8f85-da8fa2ff942b" (UID: "40c94c41-2c17-4cb0-8f85-da8fa2ff942b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.217017 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "40c94c41-2c17-4cb0-8f85-da8fa2ff942b" (UID: "40c94c41-2c17-4cb0-8f85-da8fa2ff942b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.248108 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.248136 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.248146 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw96d\" (UniqueName: \"kubernetes.io/projected/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-kube-api-access-qw96d\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.248156 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.248167 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40c94c41-2c17-4cb0-8f85-da8fa2ff942b-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.274917 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.299673 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0298-account-create-bt4k8"] Nov 24 00:44:17 crc kubenswrapper[4888]: W1124 00:44:17.299905 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode50db534_1d78_4650_ba32_f6c4eaedc101.slice/crio-cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d WatchSource:0}: Error finding container cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d: Status 404 returned error can't find the container with id cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.316402 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bq8p7"] Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.323555 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vm5vz"] Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.447715 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aa04-account-create-c6q7c" event={"ID":"1af2ab20-ea96-42e5-9aa1-3a319b4a5630","Type":"ContainerDied","Data":"db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.447768 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db1cd77c1e16d8820d16ef5e5aaafff69f895fb14a5c8fbbfa8258e6e1c6cc31" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.447734 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aa04-account-create-c6q7c" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.449207 4888 generic.go:334] "Generic (PLEG): container finished" podID="0faaac94-c8d8-4aa3-a321-250f7bf4c252" containerID="cea07138e0b390031e19f6d87e7d0a447aa791ec115d30341211198230033db1" exitCode=0 Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.449271 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wx4qp" event={"ID":"0faaac94-c8d8-4aa3-a321-250f7bf4c252","Type":"ContainerDied","Data":"cea07138e0b390031e19f6d87e7d0a447aa791ec115d30341211198230033db1"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.449298 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wx4qp" event={"ID":"0faaac94-c8d8-4aa3-a321-250f7bf4c252","Type":"ContainerStarted","Data":"de006ff913f95f81c123135e10bbd0659e872f25336a871a7ecefecf41ca4278"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.457166 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" event={"ID":"0051d661-e49b-4b29-8532-a975ba76ba66","Type":"ContainerStarted","Data":"12f7e20b33df610d0228664612e6dc7cdc6b25c0a636a897063facd5e4837ef9"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.460101 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" event={"ID":"e50db534-1d78-4650-ba32-f6c4eaedc101","Type":"ContainerStarted","Data":"cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.466157 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jpqvm" event={"ID":"47fc5340-6cad-4102-8d70-c57ea387f607","Type":"ContainerDied","Data":"d15e13697dcb613e4756b9cda9e2a14d47be37499c5bf605c5cf1c7c1e10539a"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.466540 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15e13697dcb613e4756b9cda9e2a14d47be37499c5bf605c5cf1c7c1e10539a" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.466178 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jpqvm" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.468332 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bq8p7" event={"ID":"1eaecd0e-0a6f-45d1-95da-c6ed86192911","Type":"ContainerStarted","Data":"6716693ced99f4eb9fd680aa6efe5b4a683762bec62ec31f1b41dc22ee140c31"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.470786 4888 generic.go:334] "Generic (PLEG): container finished" podID="4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" containerID="bc40d1929d02ba606926d22c27ab8db80cea6f31bc956924dd08dbfbb2eb2919" exitCode=0 Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.470884 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3726-account-create-7nrm5" event={"ID":"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9","Type":"ContainerDied","Data":"bc40d1929d02ba606926d22c27ab8db80cea6f31bc956924dd08dbfbb2eb2919"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.470949 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3726-account-create-7nrm5" event={"ID":"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9","Type":"ContainerStarted","Data":"c7953623344ae570dbe04261456c19f31600a4af728a4f32dda3125790d64270"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.479698 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerStarted","Data":"7cd53609af3dc2fa10a4c55cba95b410578e80a66caed31d05efe6587a89bf32"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.483349 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" event={"ID":"40c94c41-2c17-4cb0-8f85-da8fa2ff942b","Type":"ContainerDied","Data":"63b7d95f886737acfa8eab0a45584d8ee7ef01d6e0383dec0fdf462d93c446dc"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.483444 4888 scope.go:117] "RemoveContainer" containerID="a0a02c64906e577a1c39a733e949ec7939c76cec1f1699748be4e19893560327" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.483386 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hbcnd" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.486392 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m5bkb" event={"ID":"cc8a2262-2ebb-4ced-bec2-2760ceb19595","Type":"ContainerStarted","Data":"51984e4816ab56d7d3e5592327c9fb4ab644caebee80c61094f46e00ca5cb00a"} Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.490295 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" podStartSLOduration=5.490282531 podStartE2EDuration="5.490282531s" podCreationTimestamp="2025-11-24 00:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:44:17.486043702 +0000 UTC m=+1160.068727746" watchObservedRunningTime="2025-11-24 00:44:17.490282531 +0000 UTC m=+1160.072966575" Nov 24 00:44:17 crc kubenswrapper[4888]: W1124 00:44:17.516167 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55d0fd22_3184_4fd6_8624_cbe4c1ce970c.slice/crio-1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d WatchSource:0}: Error finding container 1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d: Status 404 returned error can't find the container with id 1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.521360 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7vxql"] Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.596727 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be14-account-create-jqd96"] Nov 24 00:44:17 crc kubenswrapper[4888]: W1124 00:44:17.605243 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7525a91c_0d49_42bf_b697_3bc21bf8769f.slice/crio-cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d WatchSource:0}: Error finding container cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d: Status 404 returned error can't find the container with id cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.928666 4888 scope.go:117] "RemoveContainer" containerID="4659b25c657f59d8328b3cfca7d066ab83a2dcd32f12a5e7d6db5473852fc206" Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.952856 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:17 crc kubenswrapper[4888]: I1124 00:44:17.962387 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hbcnd"] Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.064625 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:18 crc kubenswrapper[4888]: E1124 00:44:18.064868 4888 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 00:44:18 crc kubenswrapper[4888]: E1124 00:44:18.064905 4888 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 00:44:18 crc kubenswrapper[4888]: E1124 00:44:18.064968 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift podName:9f931286-c309-4f90-8c3e-7f2b3682e68c nodeName:}" failed. No retries permitted until 2025-11-24 00:44:22.064946461 +0000 UTC m=+1164.647630505 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift") pod "swift-storage-0" (UID: "9f931286-c309-4f90-8c3e-7f2b3682e68c") : configmap "swift-ring-files" not found Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.260923 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" path="/var/lib/kubelet/pods/40c94c41-2c17-4cb0-8f85-da8fa2ff942b/volumes" Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.502252 4888 generic.go:334] "Generic (PLEG): container finished" podID="55d0fd22-3184-4fd6-8624-cbe4c1ce970c" containerID="68cd37fe99d620c36b75d67ae7670ea2f45f4aaf8b1bdab7191ff847cdd6ad35" exitCode=0 Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.502292 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7vxql" event={"ID":"55d0fd22-3184-4fd6-8624-cbe4c1ce970c","Type":"ContainerDied","Data":"68cd37fe99d620c36b75d67ae7670ea2f45f4aaf8b1bdab7191ff847cdd6ad35"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.502563 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7vxql" event={"ID":"55d0fd22-3184-4fd6-8624-cbe4c1ce970c","Type":"ContainerStarted","Data":"1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.506624 4888 generic.go:334] "Generic (PLEG): container finished" podID="7525a91c-0d49-42bf-b697-3bc21bf8769f" containerID="65b776c2e8937ce2b76d404fc17bb108320f618ade0ccf8082b9ee1597a23294" exitCode=0 Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.506742 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be14-account-create-jqd96" event={"ID":"7525a91c-0d49-42bf-b697-3bc21bf8769f","Type":"ContainerDied","Data":"65b776c2e8937ce2b76d404fc17bb108320f618ade0ccf8082b9ee1597a23294"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.506798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be14-account-create-jqd96" event={"ID":"7525a91c-0d49-42bf-b697-3bc21bf8769f","Type":"ContainerStarted","Data":"cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.508570 4888 generic.go:334] "Generic (PLEG): container finished" podID="0051d661-e49b-4b29-8532-a975ba76ba66" containerID="5909d20f57ef12fa2c07fe6d1e62465c189ae01a74f7deaadd8dc46213291040" exitCode=0 Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.508608 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" event={"ID":"0051d661-e49b-4b29-8532-a975ba76ba66","Type":"ContainerDied","Data":"5909d20f57ef12fa2c07fe6d1e62465c189ae01a74f7deaadd8dc46213291040"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.514956 4888 generic.go:334] "Generic (PLEG): container finished" podID="e50db534-1d78-4650-ba32-f6c4eaedc101" containerID="28ae2080e28b50f3f96990a49d9ba8d389a9e74f704786ddeac56f1ceb654c6a" exitCode=0 Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.515039 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" event={"ID":"e50db534-1d78-4650-ba32-f6c4eaedc101","Type":"ContainerDied","Data":"28ae2080e28b50f3f96990a49d9ba8d389a9e74f704786ddeac56f1ceb654c6a"} Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.519304 4888 generic.go:334] "Generic (PLEG): container finished" podID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerID="d62cb800841d9a442ffe7ef47e01010a34dff0cc20cc27f85cdf489cab83d217" exitCode=0 Nov 24 00:44:18 crc kubenswrapper[4888]: I1124 00:44:18.519379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m5bkb" event={"ID":"cc8a2262-2ebb-4ced-bec2-2760ceb19595","Type":"ContainerDied","Data":"d62cb800841d9a442ffe7ef47e01010a34dff0cc20cc27f85cdf489cab83d217"} Nov 24 00:44:18 crc kubenswrapper[4888]: E1124 00:44:18.759571 4888 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.409719 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.421858 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.502731 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc6jf\" (UniqueName: \"kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf\") pod \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.502793 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts\") pod \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.502912 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts\") pod \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\" (UID: \"0faaac94-c8d8-4aa3-a321-250f7bf4c252\") " Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.502948 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hrc2\" (UniqueName: \"kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2\") pod \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\" (UID: \"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9\") " Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.503660 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" (UID: "4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.503690 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0faaac94-c8d8-4aa3-a321-250f7bf4c252" (UID: "0faaac94-c8d8-4aa3-a321-250f7bf4c252"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.529715 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3726-account-create-7nrm5" event={"ID":"4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9","Type":"ContainerDied","Data":"c7953623344ae570dbe04261456c19f31600a4af728a4f32dda3125790d64270"} Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.529776 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7953623344ae570dbe04261456c19f31600a4af728a4f32dda3125790d64270" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.529733 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3726-account-create-7nrm5" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.532206 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wx4qp" event={"ID":"0faaac94-c8d8-4aa3-a321-250f7bf4c252","Type":"ContainerDied","Data":"de006ff913f95f81c123135e10bbd0659e872f25336a871a7ecefecf41ca4278"} Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.532357 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de006ff913f95f81c123135e10bbd0659e872f25336a871a7ecefecf41ca4278" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.532451 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wx4qp" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.605473 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0faaac94-c8d8-4aa3-a321-250f7bf4c252-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.605756 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.645721 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf" (OuterVolumeSpecName: "kube-api-access-sc6jf") pod "0faaac94-c8d8-4aa3-a321-250f7bf4c252" (UID: "0faaac94-c8d8-4aa3-a321-250f7bf4c252"). InnerVolumeSpecName "kube-api-access-sc6jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.662056 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2" (OuterVolumeSpecName: "kube-api-access-7hrc2") pod "4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" (UID: "4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9"). InnerVolumeSpecName "kube-api-access-7hrc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.707865 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hrc2\" (UniqueName: \"kubernetes.io/projected/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9-kube-api-access-7hrc2\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:19 crc kubenswrapper[4888]: I1124 00:44:19.707899 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc6jf\" (UniqueName: \"kubernetes.io/projected/0faaac94-c8d8-4aa3-a321-250f7bf4c252-kube-api-access-sc6jf\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:20 crc kubenswrapper[4888]: I1124 00:44:20.216063 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-597bdc9c68-fqxrz" podUID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" containerName="console" containerID="cri-o://d1e6ca6ed1d93ac947861b41f240dba5684f9128dd4aa37488943c91b6f6e07c" gracePeriod=15 Nov 24 00:44:20 crc kubenswrapper[4888]: I1124 00:44:20.541420 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-597bdc9c68-fqxrz_d41a16d1-0335-4e9e-b64d-f21337fcbe82/console/0.log" Nov 24 00:44:20 crc kubenswrapper[4888]: I1124 00:44:20.541469 4888 generic.go:334] "Generic (PLEG): container finished" podID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" containerID="d1e6ca6ed1d93ac947861b41f240dba5684f9128dd4aa37488943c91b6f6e07c" exitCode=2 Nov 24 00:44:20 crc kubenswrapper[4888]: I1124 00:44:20.541525 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-597bdc9c68-fqxrz" event={"ID":"d41a16d1-0335-4e9e-b64d-f21337fcbe82","Type":"ContainerDied","Data":"d1e6ca6ed1d93ac947861b41f240dba5684f9128dd4aa37488943c91b6f6e07c"} Nov 24 00:44:20 crc kubenswrapper[4888]: I1124 00:44:20.543508 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerStarted","Data":"048deb521f038ca4d3c4528f84ba11e89828d478dd94b7344e72c789c26e78c6"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.815324 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.913296 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.919947 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.930542 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4svl\" (UniqueName: \"kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl\") pod \"7525a91c-0d49-42bf-b697-3bc21bf8769f\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.930646 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts\") pod \"7525a91c-0d49-42bf-b697-3bc21bf8769f\" (UID: \"7525a91c-0d49-42bf-b697-3bc21bf8769f\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.931451 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7525a91c-0d49-42bf-b697-3bc21bf8769f" (UID: "7525a91c-0d49-42bf-b697-3bc21bf8769f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.938850 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl" (OuterVolumeSpecName: "kube-api-access-z4svl") pod "7525a91c-0d49-42bf-b697-3bc21bf8769f" (UID: "7525a91c-0d49-42bf-b697-3bc21bf8769f"). InnerVolumeSpecName "kube-api-access-z4svl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.949933 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7vxql" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.957693 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-597bdc9c68-fqxrz_d41a16d1-0335-4e9e-b64d-f21337fcbe82/console/0.log" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:20.957743 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.033843 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.034979 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.034862 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035045 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035071 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts\") pod \"0051d661-e49b-4b29-8532-a975ba76ba66\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035118 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts\") pod \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035161 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035180 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhdcb\" (UniqueName: \"kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb\") pod \"e50db534-1d78-4650-ba32-f6c4eaedc101\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035216 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts\") pod \"e50db534-1d78-4650-ba32-f6c4eaedc101\" (UID: \"e50db534-1d78-4650-ba32-f6c4eaedc101\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035249 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fp28\" (UniqueName: \"kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28\") pod \"0051d661-e49b-4b29-8532-a975ba76ba66\" (UID: \"0051d661-e49b-4b29-8532-a975ba76ba66\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035284 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035768 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035323 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035900 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7\") pod \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\" (UID: \"d41a16d1-0335-4e9e-b64d-f21337fcbe82\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.035963 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpb6l\" (UniqueName: \"kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l\") pod \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\" (UID: \"55d0fd22-3184-4fd6-8624-cbe4c1ce970c\") " Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.036775 4888 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.036790 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4svl\" (UniqueName: \"kubernetes.io/projected/7525a91c-0d49-42bf-b697-3bc21bf8769f-kube-api-access-z4svl\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.036800 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7525a91c-0d49-42bf-b697-3bc21bf8769f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.036825 4888 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.038135 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config" (OuterVolumeSpecName: "console-config") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.039140 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55d0fd22-3184-4fd6-8624-cbe4c1ce970c" (UID: "55d0fd22-3184-4fd6-8624-cbe4c1ce970c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.039572 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca" (OuterVolumeSpecName: "service-ca") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.039677 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0051d661-e49b-4b29-8532-a975ba76ba66" (UID: "0051d661-e49b-4b29-8532-a975ba76ba66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.039746 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e50db534-1d78-4650-ba32-f6c4eaedc101" (UID: "e50db534-1d78-4650-ba32-f6c4eaedc101"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.041066 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.041118 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28" (OuterVolumeSpecName: "kube-api-access-2fp28") pod "0051d661-e49b-4b29-8532-a975ba76ba66" (UID: "0051d661-e49b-4b29-8532-a975ba76ba66"). InnerVolumeSpecName "kube-api-access-2fp28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.041167 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7" (OuterVolumeSpecName: "kube-api-access-6w2g7") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "kube-api-access-6w2g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.042277 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d41a16d1-0335-4e9e-b64d-f21337fcbe82" (UID: "d41a16d1-0335-4e9e-b64d-f21337fcbe82"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.043290 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l" (OuterVolumeSpecName: "kube-api-access-wpb6l") pod "55d0fd22-3184-4fd6-8624-cbe4c1ce970c" (UID: "55d0fd22-3184-4fd6-8624-cbe4c1ce970c"). InnerVolumeSpecName "kube-api-access-wpb6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.044736 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb" (OuterVolumeSpecName: "kube-api-access-hhdcb") pod "e50db534-1d78-4650-ba32-f6c4eaedc101" (UID: "e50db534-1d78-4650-ba32-f6c4eaedc101"). InnerVolumeSpecName "kube-api-access-hhdcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140857 4888 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140898 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0051d661-e49b-4b29-8532-a975ba76ba66-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140913 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140929 4888 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140944 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhdcb\" (UniqueName: \"kubernetes.io/projected/e50db534-1d78-4650-ba32-f6c4eaedc101-kube-api-access-hhdcb\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140955 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e50db534-1d78-4650-ba32-f6c4eaedc101-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140967 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fp28\" (UniqueName: \"kubernetes.io/projected/0051d661-e49b-4b29-8532-a975ba76ba66-kube-api-access-2fp28\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140980 4888 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.140989 4888 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d41a16d1-0335-4e9e-b64d-f21337fcbe82-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.141004 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/d41a16d1-0335-4e9e-b64d-f21337fcbe82-kube-api-access-6w2g7\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.141016 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpb6l\" (UniqueName: \"kubernetes.io/projected/55d0fd22-3184-4fd6-8624-cbe4c1ce970c-kube-api-access-wpb6l\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.559406 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" event={"ID":"0051d661-e49b-4b29-8532-a975ba76ba66","Type":"ContainerDied","Data":"12f7e20b33df610d0228664612e6dc7cdc6b25c0a636a897063facd5e4837ef9"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.559465 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12f7e20b33df610d0228664612e6dc7cdc6b25c0a636a897063facd5e4837ef9" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.559547 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0298-account-create-bt4k8" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.578907 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" event={"ID":"e50db534-1d78-4650-ba32-f6c4eaedc101","Type":"ContainerDied","Data":"cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.578968 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdc49d61fc2025ae1d18659578a34e73dd4c1c55452edc9cc822fcce9511e04d" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.579045 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vm5vz" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.587676 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m5bkb" event={"ID":"cc8a2262-2ebb-4ced-bec2-2760ceb19595","Type":"ContainerStarted","Data":"e8820fc33a81d4ead6e9168c9ef6d26c2debe7de010bd4fac2a2310daba66c49"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.587798 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.589167 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bq8p7" event={"ID":"1eaecd0e-0a6f-45d1-95da-c6ed86192911","Type":"ContainerStarted","Data":"9e0e618eed9eab5b148006682e19a7e24a1bec47bc8bb6b0ebe3f00302e34b68"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.590490 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7vxql" event={"ID":"55d0fd22-3184-4fd6-8624-cbe4c1ce970c","Type":"ContainerDied","Data":"1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.590517 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1559f023d8822fc574023ba84e8a6893a9baa2e8fff4e848f48cbdf940d0736d" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.590565 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7vxql" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.598939 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be14-account-create-jqd96" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.598950 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be14-account-create-jqd96" event={"ID":"7525a91c-0d49-42bf-b697-3bc21bf8769f","Type":"ContainerDied","Data":"cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.599003 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd0d99dfe2be6cc86c74bf82a3abfea121ccbb9247c10a5ff5faf2ef8dde356d" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.609309 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-m5bkb" podStartSLOduration=8.60911217 podStartE2EDuration="8.60911217s" podCreationTimestamp="2025-11-24 00:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:44:21.605264102 +0000 UTC m=+1164.187948156" watchObservedRunningTime="2025-11-24 00:44:21.60911217 +0000 UTC m=+1164.191796214" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.617091 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-597bdc9c68-fqxrz_d41a16d1-0335-4e9e-b64d-f21337fcbe82/console/0.log" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.617259 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-597bdc9c68-fqxrz" event={"ID":"d41a16d1-0335-4e9e-b64d-f21337fcbe82","Type":"ContainerDied","Data":"8b0f7c68853638d22eeb419c2d643928ca73ed792bf3f66172d1566c64b88459"} Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.617490 4888 scope.go:117] "RemoveContainer" containerID="d1e6ca6ed1d93ac947861b41f240dba5684f9128dd4aa37488943c91b6f6e07c" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.617574 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-597bdc9c68-fqxrz" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.622681 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bq8p7" podStartSLOduration=4.258116165 podStartE2EDuration="7.622664959s" podCreationTimestamp="2025-11-24 00:44:14 +0000 UTC" firstStartedPulling="2025-11-24 00:44:17.306169632 +0000 UTC m=+1159.888853686" lastFinishedPulling="2025-11-24 00:44:20.670718436 +0000 UTC m=+1163.253402480" observedRunningTime="2025-11-24 00:44:21.61948829 +0000 UTC m=+1164.202172334" watchObservedRunningTime="2025-11-24 00:44:21.622664959 +0000 UTC m=+1164.205348993" Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.701173 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:44:21 crc kubenswrapper[4888]: I1124 00:44:21.712238 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-597bdc9c68-fqxrz"] Nov 24 00:44:22 crc kubenswrapper[4888]: I1124 00:44:22.074698 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 00:44:22 crc kubenswrapper[4888]: I1124 00:44:22.164370 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:22 crc kubenswrapper[4888]: E1124 00:44:22.164990 4888 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 00:44:22 crc kubenswrapper[4888]: E1124 00:44:22.165027 4888 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 00:44:22 crc kubenswrapper[4888]: E1124 00:44:22.165107 4888 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift podName:9f931286-c309-4f90-8c3e-7f2b3682e68c nodeName:}" failed. No retries permitted until 2025-11-24 00:44:30.165087258 +0000 UTC m=+1172.747771312 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift") pod "swift-storage-0" (UID: "9f931286-c309-4f90-8c3e-7f2b3682e68c") : configmap "swift-ring-files" not found Nov 24 00:44:22 crc kubenswrapper[4888]: I1124 00:44:22.273060 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" path="/var/lib/kubelet/pods/d41a16d1-0335-4e9e-b64d-f21337fcbe82/volumes" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.291722 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-4s525"] Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292445 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="init" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292461 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="init" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292477 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292485 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292500 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="dnsmasq-dns" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292511 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="dnsmasq-dns" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292529 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7525a91c-0d49-42bf-b697-3bc21bf8769f" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292539 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7525a91c-0d49-42bf-b697-3bc21bf8769f" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292553 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0faaac94-c8d8-4aa3-a321-250f7bf4c252" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292563 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0faaac94-c8d8-4aa3-a321-250f7bf4c252" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292583 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0051d661-e49b-4b29-8532-a975ba76ba66" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292592 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0051d661-e49b-4b29-8532-a975ba76ba66" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292611 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d0fd22-3184-4fd6-8624-cbe4c1ce970c" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292618 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d0fd22-3184-4fd6-8624-cbe4c1ce970c" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292630 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50db534-1d78-4650-ba32-f6c4eaedc101" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292637 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50db534-1d78-4650-ba32-f6c4eaedc101" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292646 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fc5340-6cad-4102-8d70-c57ea387f607" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292654 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fc5340-6cad-4102-8d70-c57ea387f607" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292675 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" containerName="console" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292682 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" containerName="console" Nov 24 00:44:23 crc kubenswrapper[4888]: E1124 00:44:23.292695 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af2ab20-ea96-42e5-9aa1-3a319b4a5630" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292703 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af2ab20-ea96-42e5-9aa1-3a319b4a5630" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292937 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="47fc5340-6cad-4102-8d70-c57ea387f607" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292954 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c94c41-2c17-4cb0-8f85-da8fa2ff942b" containerName="dnsmasq-dns" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292969 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d0fd22-3184-4fd6-8624-cbe4c1ce970c" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292981 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7525a91c-0d49-42bf-b697-3bc21bf8769f" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.292993 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0faaac94-c8d8-4aa3-a321-250f7bf4c252" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293012 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50db534-1d78-4650-ba32-f6c4eaedc101" containerName="mariadb-database-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293025 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0051d661-e49b-4b29-8532-a975ba76ba66" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293040 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="d41a16d1-0335-4e9e-b64d-f21337fcbe82" containerName="console" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293057 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af2ab20-ea96-42e5-9aa1-3a319b4a5630" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293071 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" containerName="mariadb-account-create" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.293880 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.300891 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-4s525"] Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.387876 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.388000 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkrjm\" (UniqueName: \"kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.489667 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.490441 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.490597 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkrjm\" (UniqueName: \"kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.504607 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-a03d-account-create-2tm5v"] Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.506332 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.508372 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.515570 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkrjm\" (UniqueName: \"kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm\") pod \"mysqld-exporter-openstack-cell1-db-create-4s525\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.534392 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-a03d-account-create-2tm5v"] Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.592325 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp8vc\" (UniqueName: \"kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.592509 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.630476 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.694711 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.694849 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp8vc\" (UniqueName: \"kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.695838 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.713936 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp8vc\" (UniqueName: \"kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc\") pod \"mysqld-exporter-a03d-account-create-2tm5v\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:23 crc kubenswrapper[4888]: I1124 00:44:23.871098 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:24 crc kubenswrapper[4888]: I1124 00:44:24.664615 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerStarted","Data":"0e67210a51f27089d628d91472bf5c892b6ead261c2d8fadf64f522489864e05"} Nov 24 00:44:24 crc kubenswrapper[4888]: I1124 00:44:24.689212 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=12.802258051999999 podStartE2EDuration="51.689196639s" podCreationTimestamp="2025-11-24 00:43:33 +0000 UTC" firstStartedPulling="2025-11-24 00:43:45.434384674 +0000 UTC m=+1128.017068708" lastFinishedPulling="2025-11-24 00:44:24.321323251 +0000 UTC m=+1166.904007295" observedRunningTime="2025-11-24 00:44:24.688614963 +0000 UTC m=+1167.271299007" watchObservedRunningTime="2025-11-24 00:44:24.689196639 +0000 UTC m=+1167.271880683" Nov 24 00:44:24 crc kubenswrapper[4888]: I1124 00:44:24.814332 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-4s525"] Nov 24 00:44:24 crc kubenswrapper[4888]: I1124 00:44:24.865040 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-a03d-account-create-2tm5v"] Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.695805 4888 generic.go:334] "Generic (PLEG): container finished" podID="f9e43b89-661f-439f-8996-e5f0501657b7" containerID="fa8055c93cd089d42c5250a5fc961c7aeac793004c67ba37727fb561c44fc754" exitCode=0 Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.696110 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" event={"ID":"f9e43b89-661f-439f-8996-e5f0501657b7","Type":"ContainerDied","Data":"fa8055c93cd089d42c5250a5fc961c7aeac793004c67ba37727fb561c44fc754"} Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.696136 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" event={"ID":"f9e43b89-661f-439f-8996-e5f0501657b7","Type":"ContainerStarted","Data":"74e2b9ee4fbf53516375da8ad0f907b2334472baba9e1f0501feac91eaa2e4b7"} Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.697933 4888 generic.go:334] "Generic (PLEG): container finished" podID="32a98ed1-1482-4368-8d18-8ae005b81b6b" containerID="74f826d0db9325ba785288ad9b49740371cc443fc4c26b14aa8383326770fe40" exitCode=0 Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.698761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" event={"ID":"32a98ed1-1482-4368-8d18-8ae005b81b6b","Type":"ContainerDied","Data":"74f826d0db9325ba785288ad9b49740371cc443fc4c26b14aa8383326770fe40"} Nov 24 00:44:25 crc kubenswrapper[4888]: I1124 00:44:25.698782 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" event={"ID":"32a98ed1-1482-4368-8d18-8ae005b81b6b","Type":"ContainerStarted","Data":"6aae80aa87cc7a4724504ad0b7c1a0a9dac18c0a5c8073c719f0c7a9b77a3102"} Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.745928 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-72247"] Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.747658 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.751930 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.752998 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5v4qp" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.768382 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-72247"] Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.871409 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.871526 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8khk\" (UniqueName: \"kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.871583 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.871635 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.973068 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.973152 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.973264 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.973350 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8khk\" (UniqueName: \"kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.982410 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.982554 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:26 crc kubenswrapper[4888]: I1124 00:44:26.987545 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:26.999945 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8khk\" (UniqueName: \"kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk\") pod \"glance-db-sync-72247\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " pod="openstack/glance-db-sync-72247" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.076630 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72247" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.230511 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.239886 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.379539 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts\") pod \"32a98ed1-1482-4368-8d18-8ae005b81b6b\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.379605 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp8vc\" (UniqueName: \"kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc\") pod \"32a98ed1-1482-4368-8d18-8ae005b81b6b\" (UID: \"32a98ed1-1482-4368-8d18-8ae005b81b6b\") " Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.379731 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts\") pod \"f9e43b89-661f-439f-8996-e5f0501657b7\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.379966 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkrjm\" (UniqueName: \"kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm\") pod \"f9e43b89-661f-439f-8996-e5f0501657b7\" (UID: \"f9e43b89-661f-439f-8996-e5f0501657b7\") " Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.380410 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32a98ed1-1482-4368-8d18-8ae005b81b6b" (UID: "32a98ed1-1482-4368-8d18-8ae005b81b6b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.380515 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9e43b89-661f-439f-8996-e5f0501657b7" (UID: "f9e43b89-661f-439f-8996-e5f0501657b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.380531 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a98ed1-1482-4368-8d18-8ae005b81b6b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.385015 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc" (OuterVolumeSpecName: "kube-api-access-jp8vc") pod "32a98ed1-1482-4368-8d18-8ae005b81b6b" (UID: "32a98ed1-1482-4368-8d18-8ae005b81b6b"). InnerVolumeSpecName "kube-api-access-jp8vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.385851 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm" (OuterVolumeSpecName: "kube-api-access-kkrjm") pod "f9e43b89-661f-439f-8996-e5f0501657b7" (UID: "f9e43b89-661f-439f-8996-e5f0501657b7"). InnerVolumeSpecName "kube-api-access-kkrjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.482120 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp8vc\" (UniqueName: \"kubernetes.io/projected/32a98ed1-1482-4368-8d18-8ae005b81b6b-kube-api-access-jp8vc\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.482155 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e43b89-661f-439f-8996-e5f0501657b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.482165 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkrjm\" (UniqueName: \"kubernetes.io/projected/f9e43b89-661f-439f-8996-e5f0501657b7-kube-api-access-kkrjm\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.630121 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-72247"] Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.719464 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" event={"ID":"32a98ed1-1482-4368-8d18-8ae005b81b6b","Type":"ContainerDied","Data":"6aae80aa87cc7a4724504ad0b7c1a0a9dac18c0a5c8073c719f0c7a9b77a3102"} Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.719514 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aae80aa87cc7a4724504ad0b7c1a0a9dac18c0a5c8073c719f0c7a9b77a3102" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.719480 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-a03d-account-create-2tm5v" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.721351 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" event={"ID":"f9e43b89-661f-439f-8996-e5f0501657b7","Type":"ContainerDied","Data":"74e2b9ee4fbf53516375da8ad0f907b2334472baba9e1f0501feac91eaa2e4b7"} Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.721384 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74e2b9ee4fbf53516375da8ad0f907b2334472baba9e1f0501feac91eaa2e4b7" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.721431 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-4s525" Nov 24 00:44:27 crc kubenswrapper[4888]: I1124 00:44:27.722417 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72247" event={"ID":"4ff49e8b-d3e9-453a-8adb-858fba7c7418","Type":"ContainerStarted","Data":"b40bff3c24f6637ea71b9533fe17c85b3e96d621d40eadb243352fcc0c79657b"} Nov 24 00:44:28 crc kubenswrapper[4888]: E1124 00:44:28.063110 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eaecd0e_0a6f_45d1_95da_c6ed86192911.slice/crio-9e0e618eed9eab5b148006682e19a7e24a1bec47bc8bb6b0ebe3f00302e34b68.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eaecd0e_0a6f_45d1_95da_c6ed86192911.slice/crio-conmon-9e0e618eed9eab5b148006682e19a7e24a1bec47bc8bb6b0ebe3f00302e34b68.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.494454 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.559903 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.560156 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="dnsmasq-dns" containerID="cri-o://3d48bbdf118735c7220919ce837f3123047a91dd1af5dbde9639dbcbb1897d7b" gracePeriod=10 Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.673444 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:44:28 crc kubenswrapper[4888]: E1124 00:44:28.673826 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e43b89-661f-439f-8996-e5f0501657b7" containerName="mariadb-database-create" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.673841 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e43b89-661f-439f-8996-e5f0501657b7" containerName="mariadb-database-create" Nov 24 00:44:28 crc kubenswrapper[4888]: E1124 00:44:28.673865 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a98ed1-1482-4368-8d18-8ae005b81b6b" containerName="mariadb-account-create" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.673871 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a98ed1-1482-4368-8d18-8ae005b81b6b" containerName="mariadb-account-create" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.674021 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e43b89-661f-439f-8996-e5f0501657b7" containerName="mariadb-database-create" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.674048 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a98ed1-1482-4368-8d18-8ae005b81b6b" containerName="mariadb-account-create" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.674614 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.677623 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.696486 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.753141 4888 generic.go:334] "Generic (PLEG): container finished" podID="1eaecd0e-0a6f-45d1-95da-c6ed86192911" containerID="9e0e618eed9eab5b148006682e19a7e24a1bec47bc8bb6b0ebe3f00302e34b68" exitCode=0 Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.753217 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bq8p7" event={"ID":"1eaecd0e-0a6f-45d1-95da-c6ed86192911","Type":"ContainerDied","Data":"9e0e618eed9eab5b148006682e19a7e24a1bec47bc8bb6b0ebe3f00302e34b68"} Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.755792 4888 generic.go:334] "Generic (PLEG): container finished" podID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerID="3d48bbdf118735c7220919ce837f3123047a91dd1af5dbde9639dbcbb1897d7b" exitCode=0 Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.755855 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" event={"ID":"6df49d3e-c85e-4969-adf0-9aee1f672f4a","Type":"ContainerDied","Data":"3d48bbdf118735c7220919ce837f3123047a91dd1af5dbde9639dbcbb1897d7b"} Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.806507 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzw7v\" (UniqueName: \"kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.806719 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.806752 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.908405 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.908448 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.908513 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzw7v\" (UniqueName: \"kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.920772 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.937688 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:28 crc kubenswrapper[4888]: I1124 00:44:28.942371 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzw7v\" (UniqueName: \"kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v\") pod \"mysqld-exporter-0\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " pod="openstack/mysqld-exporter-0" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.058570 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.172939 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.315537 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc\") pod \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.315984 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtctx\" (UniqueName: \"kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx\") pod \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.316025 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config\") pod \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\" (UID: \"6df49d3e-c85e-4969-adf0-9aee1f672f4a\") " Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.323543 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx" (OuterVolumeSpecName: "kube-api-access-wtctx") pod "6df49d3e-c85e-4969-adf0-9aee1f672f4a" (UID: "6df49d3e-c85e-4969-adf0-9aee1f672f4a"). InnerVolumeSpecName "kube-api-access-wtctx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.364583 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config" (OuterVolumeSpecName: "config") pod "6df49d3e-c85e-4969-adf0-9aee1f672f4a" (UID: "6df49d3e-c85e-4969-adf0-9aee1f672f4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.378897 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.388106 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6df49d3e-c85e-4969-adf0-9aee1f672f4a" (UID: "6df49d3e-c85e-4969-adf0-9aee1f672f4a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.417924 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.417971 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtctx\" (UniqueName: \"kubernetes.io/projected/6df49d3e-c85e-4969-adf0-9aee1f672f4a-kube-api-access-wtctx\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.417987 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df49d3e-c85e-4969-adf0-9aee1f672f4a-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.522722 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.765389 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" event={"ID":"6df49d3e-c85e-4969-adf0-9aee1f672f4a","Type":"ContainerDied","Data":"a431b4effe092bb901a95e439c84bfbc3295b186274dde1b0b409df6a85f7ced"} Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.765445 4888 scope.go:117] "RemoveContainer" containerID="3d48bbdf118735c7220919ce837f3123047a91dd1af5dbde9639dbcbb1897d7b" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.765490 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f8bps" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.766990 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"1565ef5a-8292-4a7e-9484-c6483c547353","Type":"ContainerStarted","Data":"c0636154a0b16eb63220790ed87311000f676387b8cf2ba48b5d397820dc0c8d"} Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.807173 4888 scope.go:117] "RemoveContainer" containerID="90b44a9ed1bd11c2486e363d5459c1546ae5820aa10a394e962786bf859500a5" Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.821865 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:44:29 crc kubenswrapper[4888]: I1124 00:44:29.833970 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f8bps"] Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.164140 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.230629 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.230780 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.230834 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.230922 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4bsl\" (UniqueName: \"kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.230952 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.231057 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.231095 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf\") pod \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\" (UID: \"1eaecd0e-0a6f-45d1-95da-c6ed86192911\") " Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.231338 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.232510 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.232522 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.238362 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl" (OuterVolumeSpecName: "kube-api-access-g4bsl") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "kube-api-access-g4bsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.242460 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9f931286-c309-4f90-8c3e-7f2b3682e68c-etc-swift\") pod \"swift-storage-0\" (UID: \"9f931286-c309-4f90-8c3e-7f2b3682e68c\") " pod="openstack/swift-storage-0" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.243893 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.258143 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" path="/var/lib/kubelet/pods/6df49d3e-c85e-4969-adf0-9aee1f672f4a/volumes" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.262939 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts" (OuterVolumeSpecName: "scripts") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.266186 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.272309 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eaecd0e-0a6f-45d1-95da-c6ed86192911" (UID: "1eaecd0e-0a6f-45d1-95da-c6ed86192911"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333317 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4bsl\" (UniqueName: \"kubernetes.io/projected/1eaecd0e-0a6f-45d1-95da-c6ed86192911-kube-api-access-g4bsl\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333558 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333567 4888 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eaecd0e-0a6f-45d1-95da-c6ed86192911-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333577 4888 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333586 4888 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333595 4888 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eaecd0e-0a6f-45d1-95da-c6ed86192911-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.333603 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eaecd0e-0a6f-45d1-95da-c6ed86192911-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.473647 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.782190 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bq8p7" Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.784496 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bq8p7" event={"ID":"1eaecd0e-0a6f-45d1-95da-c6ed86192911","Type":"ContainerDied","Data":"6716693ced99f4eb9fd680aa6efe5b4a683762bec62ec31f1b41dc22ee140c31"} Nov 24 00:44:30 crc kubenswrapper[4888]: I1124 00:44:30.784536 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6716693ced99f4eb9fd680aa6efe5b4a683762bec62ec31f1b41dc22ee140c31" Nov 24 00:44:31 crc kubenswrapper[4888]: I1124 00:44:31.670647 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 00:44:31 crc kubenswrapper[4888]: W1124 00:44:31.681998 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f931286_c309_4f90_8c3e_7f2b3682e68c.slice/crio-2a64e7cfb922d0c73b36de1b84c19f94f2cf023c4cf42951c9552d6877766beb WatchSource:0}: Error finding container 2a64e7cfb922d0c73b36de1b84c19f94f2cf023c4cf42951c9552d6877766beb: Status 404 returned error can't find the container with id 2a64e7cfb922d0c73b36de1b84c19f94f2cf023c4cf42951c9552d6877766beb Nov 24 00:44:31 crc kubenswrapper[4888]: I1124 00:44:31.813266 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"1565ef5a-8292-4a7e-9484-c6483c547353","Type":"ContainerStarted","Data":"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5"} Nov 24 00:44:31 crc kubenswrapper[4888]: I1124 00:44:31.817205 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"2a64e7cfb922d0c73b36de1b84c19f94f2cf023c4cf42951c9552d6877766beb"} Nov 24 00:44:31 crc kubenswrapper[4888]: I1124 00:44:31.843085 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.150273116 podStartE2EDuration="3.843063397s" podCreationTimestamp="2025-11-24 00:44:28 +0000 UTC" firstStartedPulling="2025-11-24 00:44:29.519781183 +0000 UTC m=+1172.102465237" lastFinishedPulling="2025-11-24 00:44:31.212571434 +0000 UTC m=+1173.795255518" observedRunningTime="2025-11-24 00:44:31.831500804 +0000 UTC m=+1174.414184858" watchObservedRunningTime="2025-11-24 00:44:31.843063397 +0000 UTC m=+1174.425747451" Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.813852 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x5tvh" podUID="053aa355-b781-49f4-8e8a-80ae64c4de62" containerName="ovn-controller" probeResult="failure" output=< Nov 24 00:44:32 crc kubenswrapper[4888]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 00:44:32 crc kubenswrapper[4888]: > Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.829225 4888 generic.go:334] "Generic (PLEG): container finished" podID="19077405-4b37-4747-a4aa-37013176475e" containerID="d49093346fd5a6787c95e5e78049f40011ec54168f4dcdf7c2296c2c4d02c337" exitCode=0 Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.829328 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerDied","Data":"d49093346fd5a6787c95e5e78049f40011ec54168f4dcdf7c2296c2c4d02c337"} Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.831229 4888 generic.go:334] "Generic (PLEG): container finished" podID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerID="77f92ef1d28bc7ff0790ff6bf0779443bddf3af848718975cf6a1958e090c6bb" exitCode=0 Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.831792 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerDied","Data":"77f92ef1d28bc7ff0790ff6bf0779443bddf3af848718975cf6a1958e090c6bb"} Nov 24 00:44:32 crc kubenswrapper[4888]: I1124 00:44:32.933212 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.848009 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerStarted","Data":"ffa92dbfd86603ffbe9031bbd1601e583f01c66d3d29826cb69f3b20bbb67ec4"} Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.848672 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.851995 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerStarted","Data":"f00812e0f8369a19c2f0bf3b384086717a2595fd39171ef4559011cc873d78e6"} Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.852216 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.856911 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"b3a7995c035b6c0b01d1b21367f60374667b7402e52906f20496e9da65153f6d"} Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.894472 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=59.60888731 podStartE2EDuration="1m7.894453157s" podCreationTimestamp="2025-11-24 00:43:26 +0000 UTC" firstStartedPulling="2025-11-24 00:43:44.576514602 +0000 UTC m=+1127.159198646" lastFinishedPulling="2025-11-24 00:43:52.862080449 +0000 UTC m=+1135.444764493" observedRunningTime="2025-11-24 00:44:33.880663841 +0000 UTC m=+1176.463347975" watchObservedRunningTime="2025-11-24 00:44:33.894453157 +0000 UTC m=+1176.477137211" Nov 24 00:44:33 crc kubenswrapper[4888]: I1124 00:44:33.931052 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.429194986 podStartE2EDuration="1m7.93102774s" podCreationTimestamp="2025-11-24 00:43:26 +0000 UTC" firstStartedPulling="2025-11-24 00:43:44.568422836 +0000 UTC m=+1127.151106890" lastFinishedPulling="2025-11-24 00:43:56.0702556 +0000 UTC m=+1138.652939644" observedRunningTime="2025-11-24 00:44:33.92855102 +0000 UTC m=+1176.511235074" watchObservedRunningTime="2025-11-24 00:44:33.93102774 +0000 UTC m=+1176.513711794" Nov 24 00:44:34 crc kubenswrapper[4888]: I1124 00:44:34.379775 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:34 crc kubenswrapper[4888]: I1124 00:44:34.386708 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:34 crc kubenswrapper[4888]: I1124 00:44:34.867547 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"1a163b83de9c1e187189b5406d71a6281b864d719836d3ea86b033073287183d"} Nov 24 00:44:34 crc kubenswrapper[4888]: I1124 00:44:34.869721 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:36 crc kubenswrapper[4888]: I1124 00:44:36.916613 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:36 crc kubenswrapper[4888]: I1124 00:44:36.917086 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="prometheus" containerID="cri-o://7cd53609af3dc2fa10a4c55cba95b410578e80a66caed31d05efe6587a89bf32" gracePeriod=600 Nov 24 00:44:36 crc kubenswrapper[4888]: I1124 00:44:36.917489 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="thanos-sidecar" containerID="cri-o://0e67210a51f27089d628d91472bf5c892b6ead261c2d8fadf64f522489864e05" gracePeriod=600 Nov 24 00:44:36 crc kubenswrapper[4888]: I1124 00:44:36.917672 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="config-reloader" containerID="cri-o://048deb521f038ca4d3c4528f84ba11e89828d478dd94b7344e72c789c26e78c6" gracePeriod=600 Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.823855 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x5tvh" podUID="053aa355-b781-49f4-8e8a-80ae64c4de62" containerName="ovn-controller" probeResult="failure" output=< Nov 24 00:44:37 crc kubenswrapper[4888]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 00:44:37 crc kubenswrapper[4888]: > Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894504 4888 generic.go:334] "Generic (PLEG): container finished" podID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerID="0e67210a51f27089d628d91472bf5c892b6ead261c2d8fadf64f522489864e05" exitCode=0 Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894533 4888 generic.go:334] "Generic (PLEG): container finished" podID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerID="048deb521f038ca4d3c4528f84ba11e89828d478dd94b7344e72c789c26e78c6" exitCode=0 Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894541 4888 generic.go:334] "Generic (PLEG): container finished" podID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerID="7cd53609af3dc2fa10a4c55cba95b410578e80a66caed31d05efe6587a89bf32" exitCode=0 Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894566 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerDied","Data":"0e67210a51f27089d628d91472bf5c892b6ead261c2d8fadf64f522489864e05"} Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894590 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerDied","Data":"048deb521f038ca4d3c4528f84ba11e89828d478dd94b7344e72c789c26e78c6"} Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.894602 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerDied","Data":"7cd53609af3dc2fa10a4c55cba95b410578e80a66caed31d05efe6587a89bf32"} Nov 24 00:44:37 crc kubenswrapper[4888]: I1124 00:44:37.912259 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-g6h6d" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.124920 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x5tvh-config-wwxvx"] Nov 24 00:44:38 crc kubenswrapper[4888]: E1124 00:44:38.125646 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eaecd0e-0a6f-45d1-95da-c6ed86192911" containerName="swift-ring-rebalance" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.125676 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eaecd0e-0a6f-45d1-95da-c6ed86192911" containerName="swift-ring-rebalance" Nov 24 00:44:38 crc kubenswrapper[4888]: E1124 00:44:38.125702 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="init" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.125707 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="init" Nov 24 00:44:38 crc kubenswrapper[4888]: E1124 00:44:38.125730 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="dnsmasq-dns" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.125749 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="dnsmasq-dns" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.125948 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df49d3e-c85e-4969-adf0-9aee1f672f4a" containerName="dnsmasq-dns" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.125988 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eaecd0e-0a6f-45d1-95da-c6ed86192911" containerName="swift-ring-rebalance" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.126782 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.129155 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.143882 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x5tvh-config-wwxvx"] Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198482 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198537 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198666 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198721 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198776 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.198803 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vnpj\" (UniqueName: \"kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.299863 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300159 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vnpj\" (UniqueName: \"kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300308 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300537 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300618 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300738 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300699 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.300859 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.301003 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.301858 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.319224 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vnpj\" (UniqueName: \"kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj\") pod \"ovn-controller-x5tvh-config-wwxvx\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:38 crc kubenswrapper[4888]: I1124 00:44:38.449559 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:39 crc kubenswrapper[4888]: I1124 00:44:39.379995 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.129:9090/-/ready\": dial tcp 10.217.0.129:9090: connect: connection refused" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.700673 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811363 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811709 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811741 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811828 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv7jq\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811927 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.811978 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.812022 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.812036 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file\") pod \"ce885b04-03a7-43c9-acd0-f26239bf4596\" (UID: \"ce885b04-03a7-43c9-acd0-f26239bf4596\") " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.813498 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.822765 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.824479 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config" (OuterVolumeSpecName: "config") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.828361 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq" (OuterVolumeSpecName: "kube-api-access-hv7jq") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "kube-api-access-hv7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.828433 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out" (OuterVolumeSpecName: "config-out") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.828678 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.841489 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x5tvh" podUID="053aa355-b781-49f4-8e8a-80ae64c4de62" containerName="ovn-controller" probeResult="failure" output=< Nov 24 00:44:42 crc kubenswrapper[4888]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 00:44:42 crc kubenswrapper[4888]: > Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.850477 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919688 4888 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919748 4888 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce885b04-03a7-43c9-acd0-f26239bf4596-config-out\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919765 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919777 4888 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919849 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") on node \"crc\" " Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919867 4888 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce885b04-03a7-43c9-acd0-f26239bf4596-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.919880 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv7jq\" (UniqueName: \"kubernetes.io/projected/ce885b04-03a7-43c9-acd0-f26239bf4596-kube-api-access-hv7jq\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.932070 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x5tvh-config-wwxvx"] Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.960525 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce885b04-03a7-43c9-acd0-f26239bf4596","Type":"ContainerDied","Data":"5b03db74d02b832cd0155e69110494340500923a78bca550c596feed69266d92"} Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.960575 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.960587 4888 scope.go:117] "RemoveContainer" containerID="0e67210a51f27089d628d91472bf5c892b6ead261c2d8fadf64f522489864e05" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.966746 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"bcd78d7f0346d3d95f40915a236854741ee7ecb3ccb344d6adb48782f798be1e"} Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.978754 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config" (OuterVolumeSpecName: "web-config") pod "ce885b04-03a7-43c9-acd0-f26239bf4596" (UID: "ce885b04-03a7-43c9-acd0-f26239bf4596"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.985255 4888 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.985377 4888 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799") on node "crc" Nov 24 00:44:42 crc kubenswrapper[4888]: I1124 00:44:42.994113 4888 scope.go:117] "RemoveContainer" containerID="048deb521f038ca4d3c4528f84ba11e89828d478dd94b7344e72c789c26e78c6" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.002567 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.020909 4888 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce885b04-03a7-43c9-acd0-f26239bf4596-web-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.020941 4888 reconciler_common.go:293] "Volume detached for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.029712 4888 scope.go:117] "RemoveContainer" containerID="7cd53609af3dc2fa10a4c55cba95b410578e80a66caed31d05efe6587a89bf32" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.064903 4888 scope.go:117] "RemoveContainer" containerID="79592bc08a061c8ba660552008cbf8fb97997b70e608e100de19683b1c0dc9a8" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.302017 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.318960 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.346976 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:43 crc kubenswrapper[4888]: E1124 00:44:43.347341 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="config-reloader" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347358 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="config-reloader" Nov 24 00:44:43 crc kubenswrapper[4888]: E1124 00:44:43.347380 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="prometheus" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347389 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="prometheus" Nov 24 00:44:43 crc kubenswrapper[4888]: E1124 00:44:43.347416 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="init-config-reloader" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347424 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="init-config-reloader" Nov 24 00:44:43 crc kubenswrapper[4888]: E1124 00:44:43.347454 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="thanos-sidecar" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347463 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="thanos-sidecar" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347640 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="config-reloader" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347663 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="thanos-sidecar" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.347675 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" containerName="prometheus" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.349271 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.352256 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.353277 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.353748 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.353777 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.353899 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.358180 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-5vsrn" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.361017 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.377463 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.435965 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436013 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n2jw\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-kube-api-access-8n2jw\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436087 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436150 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436206 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436227 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436286 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436343 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.436382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.537664 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538248 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538298 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538317 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n2jw\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-kube-api-access-8n2jw\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538377 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538421 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538449 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538471 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538490 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538510 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.538531 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.542093 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.545215 4888 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.545255 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/94ad0ac163fed009bb5498203f4a81a25a208cff03033a2bd0264ed6773db54c/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.547081 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.547855 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.549054 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.553385 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.557024 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.557869 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.558248 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.558263 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.566936 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n2jw\" (UniqueName: \"kubernetes.io/projected/7ba9e0fb-846a-49dc-bdc3-a0e201324dab-kube-api-access-8n2jw\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.595935 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b0ca225-84fb-40f6-b4fc-bc9b769c4799\") pod \"prometheus-metric-storage-0\" (UID: \"7ba9e0fb-846a-49dc-bdc3-a0e201324dab\") " pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.757778 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.977327 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"eba0abc7c73fe8b92fceb2b00a0afc03e1b4a947f18d2b6ede22f590bbfd3ca7"} Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.979185 4888 generic.go:334] "Generic (PLEG): container finished" podID="9c03e529-1c77-46cf-9a25-2a94c4da23b6" containerID="34f418c2d731b62722a2089e7af7bede3064c9909b7e01445dfcb5b698fccf88" exitCode=0 Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.979484 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x5tvh-config-wwxvx" event={"ID":"9c03e529-1c77-46cf-9a25-2a94c4da23b6","Type":"ContainerDied","Data":"34f418c2d731b62722a2089e7af7bede3064c9909b7e01445dfcb5b698fccf88"} Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.979529 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x5tvh-config-wwxvx" event={"ID":"9c03e529-1c77-46cf-9a25-2a94c4da23b6","Type":"ContainerStarted","Data":"917503b119425ccce07732e83b2edefc18a87ef8e82d918a2e0d09239ab44d45"} Nov 24 00:44:43 crc kubenswrapper[4888]: I1124 00:44:43.981388 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72247" event={"ID":"4ff49e8b-d3e9-453a-8adb-858fba7c7418","Type":"ContainerStarted","Data":"4638da5c25e3048b04ac408c4f4b75cbe01cca5fbf3d92b2cf5cbef74054986d"} Nov 24 00:44:44 crc kubenswrapper[4888]: I1124 00:44:44.011904 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-72247" podStartSLOduration=3.179208387 podStartE2EDuration="18.011749491s" podCreationTimestamp="2025-11-24 00:44:26 +0000 UTC" firstStartedPulling="2025-11-24 00:44:27.638976823 +0000 UTC m=+1170.221660867" lastFinishedPulling="2025-11-24 00:44:42.471517927 +0000 UTC m=+1185.054201971" observedRunningTime="2025-11-24 00:44:44.011253347 +0000 UTC m=+1186.593937391" watchObservedRunningTime="2025-11-24 00:44:44.011749491 +0000 UTC m=+1186.594433535" Nov 24 00:44:44 crc kubenswrapper[4888]: I1124 00:44:44.263090 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce885b04-03a7-43c9-acd0-f26239bf4596" path="/var/lib/kubelet/pods/ce885b04-03a7-43c9-acd0-f26239bf4596/volumes" Nov 24 00:44:44 crc kubenswrapper[4888]: I1124 00:44:44.264501 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 00:44:44 crc kubenswrapper[4888]: W1124 00:44:44.526129 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ba9e0fb_846a_49dc_bdc3_a0e201324dab.slice/crio-1b35a03ca1dc6e545729b0e0bbb2c273b3b52ab50421d2995a30a309c7486585 WatchSource:0}: Error finding container 1b35a03ca1dc6e545729b0e0bbb2c273b3b52ab50421d2995a30a309c7486585: Status 404 returned error can't find the container with id 1b35a03ca1dc6e545729b0e0bbb2c273b3b52ab50421d2995a30a309c7486585 Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.014725 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerStarted","Data":"1b35a03ca1dc6e545729b0e0bbb2c273b3b52ab50421d2995a30a309c7486585"} Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.025251 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"c2224c21d1986f1051d10d348bc53d85bcc8853bf363e96bf3b45decbd6a6d73"} Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.376057 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.476923 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477067 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run" (OuterVolumeSpecName: "var-run") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477102 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477365 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477413 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477462 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vnpj\" (UniqueName: \"kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477462 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477483 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn\") pod \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\" (UID: \"9c03e529-1c77-46cf-9a25-2a94c4da23b6\") " Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477866 4888 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477883 4888 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.477912 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.478617 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.478831 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts" (OuterVolumeSpecName: "scripts") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.488468 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj" (OuterVolumeSpecName: "kube-api-access-7vnpj") pod "9c03e529-1c77-46cf-9a25-2a94c4da23b6" (UID: "9c03e529-1c77-46cf-9a25-2a94c4da23b6"). InnerVolumeSpecName "kube-api-access-7vnpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.580133 4888 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.580182 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c03e529-1c77-46cf-9a25-2a94c4da23b6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.580192 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vnpj\" (UniqueName: \"kubernetes.io/projected/9c03e529-1c77-46cf-9a25-2a94c4da23b6-kube-api-access-7vnpj\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:45 crc kubenswrapper[4888]: I1124 00:44:45.580209 4888 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c03e529-1c77-46cf-9a25-2a94c4da23b6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.039712 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"230878fb8e43b271bb444686f6a11e7ee255b2c10717378f60c6416abe89abdd"} Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.039887 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"6abb4e4798ec6b4479d89f1d31dbdcb2bf99fe9641d8000bf6fffc20f93bfde6"} Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.039932 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"571fbfc568e2ac1b4010e4d6dca4c685aea135b1459a2b143c0ebdbcb26665d6"} Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.042454 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x5tvh-config-wwxvx" event={"ID":"9c03e529-1c77-46cf-9a25-2a94c4da23b6","Type":"ContainerDied","Data":"917503b119425ccce07732e83b2edefc18a87ef8e82d918a2e0d09239ab44d45"} Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.042515 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="917503b119425ccce07732e83b2edefc18a87ef8e82d918a2e0d09239ab44d45" Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.042525 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x5tvh-config-wwxvx" Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.505079 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x5tvh-config-wwxvx"] Nov 24 00:44:46 crc kubenswrapper[4888]: I1124 00:44:46.522029 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x5tvh-config-wwxvx"] Nov 24 00:44:47 crc kubenswrapper[4888]: I1124 00:44:47.628040 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 00:44:47 crc kubenswrapper[4888]: I1124 00:44:47.828422 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-x5tvh" Nov 24 00:44:47 crc kubenswrapper[4888]: I1124 00:44:47.925004 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.060936 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pn29n"] Nov 24 00:44:48 crc kubenswrapper[4888]: E1124 00:44:48.065185 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c03e529-1c77-46cf-9a25-2a94c4da23b6" containerName="ovn-config" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.065212 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c03e529-1c77-46cf-9a25-2a94c4da23b6" containerName="ovn-config" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.065411 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c03e529-1c77-46cf-9a25-2a94c4da23b6" containerName="ovn-config" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.066039 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.080564 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pn29n"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.227850 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8jsk\" (UniqueName: \"kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.227899 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.260319 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c03e529-1c77-46cf-9a25-2a94c4da23b6" path="/var/lib/kubelet/pods/9c03e529-1c77-46cf-9a25-2a94c4da23b6/volumes" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.274906 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-58e8-account-create-ths8q"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.276004 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.277470 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.290434 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-58e8-account-create-ths8q"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.329251 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8jsk\" (UniqueName: \"kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.329314 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.330207 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.352636 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xsvt9"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.358073 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.375882 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8jsk\" (UniqueName: \"kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk\") pod \"heat-db-create-pn29n\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.377628 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xsvt9"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.393714 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-792e-account-create-9xf57"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.395569 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.399048 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.420823 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-792e-account-create-9xf57"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.435885 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.436026 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d78vg\" (UniqueName: \"kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.494470 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rl79r"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.496060 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.508926 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rl79r"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.530496 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pn29n" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.533632 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-b919-account-create-pjsgg"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.534925 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.537843 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538015 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d78vg\" (UniqueName: \"kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538077 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2kfw\" (UniqueName: \"kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538201 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk6rd\" (UniqueName: \"kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538236 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538262 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.538042 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.550706 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b919-account-create-pjsgg"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.566439 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.577687 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d78vg\" (UniqueName: \"kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg\") pod \"barbican-58e8-account-create-ths8q\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.604913 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.676659 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2kfw\" (UniqueName: \"kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.676901 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk6rd\" (UniqueName: \"kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.676933 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.676984 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.677009 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffxsc\" (UniqueName: \"kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.681204 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggwpp\" (UniqueName: \"kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.681297 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.681330 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.681728 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.682038 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.706314 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-gzwq7"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.712931 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk6rd\" (UniqueName: \"kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd\") pod \"cinder-792e-account-create-9xf57\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.718520 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2kfw\" (UniqueName: \"kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw\") pod \"cinder-db-create-xsvt9\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.729795 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gzwq7"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.729935 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.735090 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2f74-account-create-kpjmw"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.742913 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.743028 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.752249 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.759938 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2f74-account-create-kpjmw"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.777396 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.782528 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggwpp\" (UniqueName: \"kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.782608 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.782846 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.783341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffxsc\" (UniqueName: \"kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.783472 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.783889 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.790727 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xqt9t"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.792113 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.795719 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.796040 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-pjwx7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.796177 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.796292 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.796384 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xqt9t"] Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.813463 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffxsc\" (UniqueName: \"kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc\") pod \"barbican-db-create-rl79r\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.815057 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggwpp\" (UniqueName: \"kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp\") pod \"heat-b919-account-create-pjsgg\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.859288 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885570 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fddgq\" (UniqueName: \"kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885623 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsgjt\" (UniqueName: \"kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885666 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885684 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnb8\" (UniqueName: \"kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885765 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885792 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.885851 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.979758 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.988941 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.988992 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.989044 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.989072 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fddgq\" (UniqueName: \"kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.989098 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgjt\" (UniqueName: \"kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.989135 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.989154 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnb8\" (UniqueName: \"kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.990249 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:48 crc kubenswrapper[4888]: I1124 00:44:48.990985 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:48.999907 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.010623 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.056341 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgjt\" (UniqueName: \"kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt\") pod \"neutron-db-create-gzwq7\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.057350 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnb8\" (UniqueName: \"kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8\") pod \"keystone-db-sync-xqt9t\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.064301 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fddgq\" (UniqueName: \"kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq\") pod \"neutron-2f74-account-create-kpjmw\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.110359 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerStarted","Data":"0a786c9ad3f7c81d1bc703c3f11316e1f170b7a6174c40ed7f8d06f881de4a88"} Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.120129 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"b6a15a8669947ba1c18fa4a2474fb7b159ff60656a78c855542c0f3167db7c7a"} Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.120178 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"0096b0966c442313b2d18c1c41b55c75533e4e6cc739bd9c4a76592d1bdda424"} Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.349213 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.393765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.418658 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.600741 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pn29n"] Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.884175 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-58e8-account-create-ths8q"] Nov 24 00:44:49 crc kubenswrapper[4888]: W1124 00:44:49.902885 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d1470a_a999_4afe_a7e4_514914054767.slice/crio-647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a WatchSource:0}: Error finding container 647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a: Status 404 returned error can't find the container with id 647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.953341 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xsvt9"] Nov 24 00:44:49 crc kubenswrapper[4888]: I1124 00:44:49.971987 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rl79r"] Nov 24 00:44:49 crc kubenswrapper[4888]: W1124 00:44:49.973883 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9609549_6956_4a00_a921_5ab8f6b0511f.slice/crio-f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76 WatchSource:0}: Error finding container f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76: Status 404 returned error can't find the container with id f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76 Nov 24 00:44:49 crc kubenswrapper[4888]: W1124 00:44:49.979037 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05103e58_6551_4742_ba88_12e3669fbee3.slice/crio-39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73 WatchSource:0}: Error finding container 39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73: Status 404 returned error can't find the container with id 39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73 Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.174656 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b919-account-create-pjsgg"] Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.178171 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xsvt9" event={"ID":"b9609549-6956-4a00-a921-5ab8f6b0511f","Type":"ContainerStarted","Data":"f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.180759 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58e8-account-create-ths8q" event={"ID":"83d1470a-a999-4afe-a7e4-514914054767","Type":"ContainerStarted","Data":"647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.192092 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xqt9t"] Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.193427 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pn29n" event={"ID":"3294b25b-9e52-43e0-ba6e-87b510640945","Type":"ContainerStarted","Data":"03bc9f2bfed81b04f3a9eda68d0ecbeca10ab83fd362b2834f7c83bdff201f7d"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.193459 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pn29n" event={"ID":"3294b25b-9e52-43e0-ba6e-87b510640945","Type":"ContainerStarted","Data":"4638d91ad4afd541727ea0f9781dcb6ae5bc625d2cd51e3aad3291f6f1d590d8"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.202935 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl79r" event={"ID":"05103e58-6551-4742-ba88-12e3669fbee3","Type":"ContainerStarted","Data":"39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.215805 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gzwq7"] Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.222280 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"9660bafb04bad71ddeec4f0282fe99901634b3e953b690c69bf3deb3c224de18"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.222328 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"6d87fa66132f19caa1d3d145bfd7b68285ffde1d8af15aa150783aecc442c371"} Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.240185 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-792e-account-create-9xf57"] Nov 24 00:44:50 crc kubenswrapper[4888]: I1124 00:44:50.300161 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2f74-account-create-kpjmw"] Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.235031 4888 generic.go:334] "Generic (PLEG): container finished" podID="5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" containerID="fa2366f9cb366682b4a51a13373c728bfdd42eb67fc98c4ad14115074093fe0f" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.235085 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gzwq7" event={"ID":"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42","Type":"ContainerDied","Data":"fa2366f9cb366682b4a51a13373c728bfdd42eb67fc98c4ad14115074093fe0f"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.235498 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gzwq7" event={"ID":"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42","Type":"ContainerStarted","Data":"89dc1484d0a826f886ac827367d9518d2cac6150a354389f2c6d638f15c0d146"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.237522 4888 generic.go:334] "Generic (PLEG): container finished" podID="47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" containerID="03708822d89411fc1b16c27682685f4c1f9a639244b751aceb5c0a145f04ea9b" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.237609 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-792e-account-create-9xf57" event={"ID":"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47","Type":"ContainerDied","Data":"03708822d89411fc1b16c27682685f4c1f9a639244b751aceb5c0a145f04ea9b"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.237631 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-792e-account-create-9xf57" event={"ID":"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47","Type":"ContainerStarted","Data":"6f59790441ca0d70be77d43aa6469505d9e95e29e25ee3816e2255e33cad4ff9"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.240066 4888 generic.go:334] "Generic (PLEG): container finished" podID="b9609549-6956-4a00-a921-5ab8f6b0511f" containerID="ff0aef7e797526e500740b14215527b7679a926ee4d9715bcc2c71bac03f5e95" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.240139 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xsvt9" event={"ID":"b9609549-6956-4a00-a921-5ab8f6b0511f","Type":"ContainerDied","Data":"ff0aef7e797526e500740b14215527b7679a926ee4d9715bcc2c71bac03f5e95"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.241292 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xqt9t" event={"ID":"c1cd14a9-9f24-436f-b96f-ea9457548844","Type":"ContainerStarted","Data":"7ac004f2d8f253c763350e89b6a8b95a6d10689f1ddd503ef31926a7eb0a4ff5"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.242540 4888 generic.go:334] "Generic (PLEG): container finished" podID="05103e58-6551-4742-ba88-12e3669fbee3" containerID="e60366f2b38f8087ea2901a120a8b3d73cb01239b68f9fe7d2e68056ed70163b" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.242589 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl79r" event={"ID":"05103e58-6551-4742-ba88-12e3669fbee3","Type":"ContainerDied","Data":"e60366f2b38f8087ea2901a120a8b3d73cb01239b68f9fe7d2e68056ed70163b"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.243685 4888 generic.go:334] "Generic (PLEG): container finished" podID="3294b25b-9e52-43e0-ba6e-87b510640945" containerID="03bc9f2bfed81b04f3a9eda68d0ecbeca10ab83fd362b2834f7c83bdff201f7d" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.243729 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pn29n" event={"ID":"3294b25b-9e52-43e0-ba6e-87b510640945","Type":"ContainerDied","Data":"03bc9f2bfed81b04f3a9eda68d0ecbeca10ab83fd362b2834f7c83bdff201f7d"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.245320 4888 generic.go:334] "Generic (PLEG): container finished" podID="a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" containerID="6f1b561bb345b91b75fea7c96d15e54444c5e6849bf646307d59120c3e07ebf2" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.245373 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b919-account-create-pjsgg" event={"ID":"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a","Type":"ContainerDied","Data":"6f1b561bb345b91b75fea7c96d15e54444c5e6849bf646307d59120c3e07ebf2"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.245397 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b919-account-create-pjsgg" event={"ID":"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a","Type":"ContainerStarted","Data":"2938ffc635916be34c21847bbbeeaa98f5e78e5ec9115016ffe579e254cc3f4e"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.254886 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"3c00b596a53007e7d485c19fec039bb1f94adce1cad18c0f0cdfd165874b47a6"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.254925 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"2f8ef5385cc11c395f1e2cf26b1fa19ace4d1e40059fe4b8261a14d3840962be"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.254936 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9f931286-c309-4f90-8c3e-7f2b3682e68c","Type":"ContainerStarted","Data":"724ae3b33ce437d487b0dc27c69e4f850cbf1b45bb4eb043e4f2a37c99a309e9"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.258670 4888 generic.go:334] "Generic (PLEG): container finished" podID="83d1470a-a999-4afe-a7e4-514914054767" containerID="acfcf3f1b38a010132c1bf6b4f04348e29b792ea91138fbe5de6d2fdcc87352a" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.258721 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58e8-account-create-ths8q" event={"ID":"83d1470a-a999-4afe-a7e4-514914054767","Type":"ContainerDied","Data":"acfcf3f1b38a010132c1bf6b4f04348e29b792ea91138fbe5de6d2fdcc87352a"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.260339 4888 generic.go:334] "Generic (PLEG): container finished" podID="61345e71-eb59-4bef-a890-65ed5a2e99b7" containerID="53e4353af5fe051831265cc7b1f8feee569e5da881744ee8e63c4ef73f5aa418" exitCode=0 Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.260366 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2f74-account-create-kpjmw" event={"ID":"61345e71-eb59-4bef-a890-65ed5a2e99b7","Type":"ContainerDied","Data":"53e4353af5fe051831265cc7b1f8feee569e5da881744ee8e63c4ef73f5aa418"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.260379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2f74-account-create-kpjmw" event={"ID":"61345e71-eb59-4bef-a890-65ed5a2e99b7","Type":"ContainerStarted","Data":"ee499d2d2b7a82f5229f66e388bebad3e4ef517504b223be0447660426627fae"} Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.347466 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=22.737830349 podStartE2EDuration="38.347449194s" podCreationTimestamp="2025-11-24 00:44:13 +0000 UTC" firstStartedPulling="2025-11-24 00:44:31.683671279 +0000 UTC m=+1174.266355323" lastFinishedPulling="2025-11-24 00:44:47.293290124 +0000 UTC m=+1189.875974168" observedRunningTime="2025-11-24 00:44:51.339866642 +0000 UTC m=+1193.922550686" watchObservedRunningTime="2025-11-24 00:44:51.347449194 +0000 UTC m=+1193.930133238" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.801285 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.803177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.809567 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.811876 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.880906 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pn29n" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.970339 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts\") pod \"3294b25b-9e52-43e0-ba6e-87b510640945\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.970571 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8jsk\" (UniqueName: \"kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk\") pod \"3294b25b-9e52-43e0-ba6e-87b510640945\" (UID: \"3294b25b-9e52-43e0-ba6e-87b510640945\") " Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.970937 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.970981 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971095 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-229wc\" (UniqueName: \"kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971123 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3294b25b-9e52-43e0-ba6e-87b510640945" (UID: "3294b25b-9e52-43e0-ba6e-87b510640945"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971163 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971284 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971325 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.971412 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3294b25b-9e52-43e0-ba6e-87b510640945-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:51 crc kubenswrapper[4888]: I1124 00:44:51.983107 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk" (OuterVolumeSpecName: "kube-api-access-b8jsk") pod "3294b25b-9e52-43e0-ba6e-87b510640945" (UID: "3294b25b-9e52-43e0-ba6e-87b510640945"). InnerVolumeSpecName "kube-api-access-b8jsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.072838 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.072915 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.072947 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.072996 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073050 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073145 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-229wc\" (UniqueName: \"kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073211 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8jsk\" (UniqueName: \"kubernetes.io/projected/3294b25b-9e52-43e0-ba6e-87b510640945-kube-api-access-b8jsk\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073864 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073869 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073921 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.073922 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.074176 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.090084 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-229wc\" (UniqueName: \"kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc\") pod \"dnsmasq-dns-764c5664d7-rfmwb\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.191348 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.275285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pn29n" event={"ID":"3294b25b-9e52-43e0-ba6e-87b510640945","Type":"ContainerDied","Data":"4638d91ad4afd541727ea0f9781dcb6ae5bc625d2cd51e3aad3291f6f1d590d8"} Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.275338 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4638d91ad4afd541727ea0f9781dcb6ae5bc625d2cd51e3aad3291f6f1d590d8" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.275337 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pn29n" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.674701 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.782364 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts\") pod \"83d1470a-a999-4afe-a7e4-514914054767\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.782591 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d78vg\" (UniqueName: \"kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg\") pod \"83d1470a-a999-4afe-a7e4-514914054767\" (UID: \"83d1470a-a999-4afe-a7e4-514914054767\") " Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.784690 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83d1470a-a999-4afe-a7e4-514914054767" (UID: "83d1470a-a999-4afe-a7e4-514914054767"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.796375 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg" (OuterVolumeSpecName: "kube-api-access-d78vg") pod "83d1470a-a999-4afe-a7e4-514914054767" (UID: "83d1470a-a999-4afe-a7e4-514914054767"). InnerVolumeSpecName "kube-api-access-d78vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.885032 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d78vg\" (UniqueName: \"kubernetes.io/projected/83d1470a-a999-4afe-a7e4-514914054767-kube-api-access-d78vg\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:52 crc kubenswrapper[4888]: I1124 00:44:52.885070 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83d1470a-a999-4afe-a7e4-514914054767-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.096630 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.102361 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.116387 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.125661 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.138006 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.178446 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189264 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fddgq\" (UniqueName: \"kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq\") pod \"61345e71-eb59-4bef-a890-65ed5a2e99b7\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189410 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts\") pod \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189538 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts\") pod \"61345e71-eb59-4bef-a890-65ed5a2e99b7\" (UID: \"61345e71-eb59-4bef-a890-65ed5a2e99b7\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189638 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts\") pod \"05103e58-6551-4742-ba88-12e3669fbee3\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189700 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggwpp\" (UniqueName: \"kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp\") pod \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\" (UID: \"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189823 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2kfw\" (UniqueName: \"kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw\") pod \"b9609549-6956-4a00-a921-5ab8f6b0511f\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189932 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts\") pod \"b9609549-6956-4a00-a921-5ab8f6b0511f\" (UID: \"b9609549-6956-4a00-a921-5ab8f6b0511f\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.190004 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffxsc\" (UniqueName: \"kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc\") pod \"05103e58-6551-4742-ba88-12e3669fbee3\" (UID: \"05103e58-6551-4742-ba88-12e3669fbee3\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.189963 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" (UID: "a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.190124 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61345e71-eb59-4bef-a890-65ed5a2e99b7" (UID: "61345e71-eb59-4bef-a890-65ed5a2e99b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.190469 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05103e58-6551-4742-ba88-12e3669fbee3" (UID: "05103e58-6551-4742-ba88-12e3669fbee3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.190591 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9609549-6956-4a00-a921-5ab8f6b0511f" (UID: "b9609549-6956-4a00-a921-5ab8f6b0511f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.195507 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc" (OuterVolumeSpecName: "kube-api-access-ffxsc") pod "05103e58-6551-4742-ba88-12e3669fbee3" (UID: "05103e58-6551-4742-ba88-12e3669fbee3"). InnerVolumeSpecName "kube-api-access-ffxsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.196986 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp" (OuterVolumeSpecName: "kube-api-access-ggwpp") pod "a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" (UID: "a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a"). InnerVolumeSpecName "kube-api-access-ggwpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.197670 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw" (OuterVolumeSpecName: "kube-api-access-j2kfw") pod "b9609549-6956-4a00-a921-5ab8f6b0511f" (UID: "b9609549-6956-4a00-a921-5ab8f6b0511f"). InnerVolumeSpecName "kube-api-access-j2kfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.207182 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq" (OuterVolumeSpecName: "kube-api-access-fddgq") pod "61345e71-eb59-4bef-a890-65ed5a2e99b7" (UID: "61345e71-eb59-4bef-a890-65ed5a2e99b7"). InnerVolumeSpecName "kube-api-access-fddgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.220042 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.291261 4888 generic.go:334] "Generic (PLEG): container finished" podID="4ff49e8b-d3e9-453a-8adb-858fba7c7418" containerID="4638da5c25e3048b04ac408c4f4b75cbe01cca5fbf3d92b2cf5cbef74054986d" exitCode=0 Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.291328 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72247" event={"ID":"4ff49e8b-d3e9-453a-8adb-858fba7c7418","Type":"ContainerDied","Data":"4638da5c25e3048b04ac408c4f4b75cbe01cca5fbf3d92b2cf5cbef74054986d"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.292759 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsgjt\" (UniqueName: \"kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt\") pod \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.292915 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk6rd\" (UniqueName: \"kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd\") pod \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.292930 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58e8-account-create-ths8q" event={"ID":"83d1470a-a999-4afe-a7e4-514914054767","Type":"ContainerDied","Data":"647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.292949 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="647f7ee69a3825948b68c21e47b1009cabeebe79beb25a0036fcddd29b49a36a" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.292956 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts\") pod \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\" (UID: \"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.293009 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts\") pod \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\" (UID: \"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47\") " Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.293016 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58e8-account-create-ths8q" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.296826 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd" (OuterVolumeSpecName: "kube-api-access-vk6rd") pod "47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" (UID: "47da0c31-1bfa-4738-a4bd-fcf0e82a3e47"). InnerVolumeSpecName "kube-api-access-vk6rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.299338 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt" (OuterVolumeSpecName: "kube-api-access-dsgjt") pod "5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" (UID: "5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42"). InnerVolumeSpecName "kube-api-access-dsgjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.300028 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" (UID: "5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.300487 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" (UID: "47da0c31-1bfa-4738-a4bd-fcf0e82a3e47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.300965 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9609549-6956-4a00-a921-5ab8f6b0511f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.300985 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffxsc\" (UniqueName: \"kubernetes.io/projected/05103e58-6551-4742-ba88-12e3669fbee3-kube-api-access-ffxsc\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.300996 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsgjt\" (UniqueName: \"kubernetes.io/projected/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-kube-api-access-dsgjt\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301005 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk6rd\" (UniqueName: \"kubernetes.io/projected/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-kube-api-access-vk6rd\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301014 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fddgq\" (UniqueName: \"kubernetes.io/projected/61345e71-eb59-4bef-a890-65ed5a2e99b7-kube-api-access-fddgq\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301023 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301031 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301062 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61345e71-eb59-4bef-a890-65ed5a2e99b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301071 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301080 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05103e58-6551-4742-ba88-12e3669fbee3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301090 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggwpp\" (UniqueName: \"kubernetes.io/projected/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a-kube-api-access-ggwpp\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301098 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2kfw\" (UniqueName: \"kubernetes.io/projected/b9609549-6956-4a00-a921-5ab8f6b0511f-kube-api-access-j2kfw\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301306 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2f74-account-create-kpjmw" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301309 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2f74-account-create-kpjmw" event={"ID":"61345e71-eb59-4bef-a890-65ed5a2e99b7","Type":"ContainerDied","Data":"ee499d2d2b7a82f5229f66e388bebad3e4ef517504b223be0447660426627fae"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.301337 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee499d2d2b7a82f5229f66e388bebad3e4ef517504b223be0447660426627fae" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.304351 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gzwq7" event={"ID":"5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42","Type":"ContainerDied","Data":"89dc1484d0a826f886ac827367d9518d2cac6150a354389f2c6d638f15c0d146"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.304376 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89dc1484d0a826f886ac827367d9518d2cac6150a354389f2c6d638f15c0d146" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.304432 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gzwq7" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.312741 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-792e-account-create-9xf57" event={"ID":"47da0c31-1bfa-4738-a4bd-fcf0e82a3e47","Type":"ContainerDied","Data":"6f59790441ca0d70be77d43aa6469505d9e95e29e25ee3816e2255e33cad4ff9"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.313189 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-792e-account-create-9xf57" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.313229 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f59790441ca0d70be77d43aa6469505d9e95e29e25ee3816e2255e33cad4ff9" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.320667 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl79r" event={"ID":"05103e58-6551-4742-ba88-12e3669fbee3","Type":"ContainerDied","Data":"39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.320713 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39e491955b12a1fe9d71e2f1bf3d291016ed416374969d1df6c25d51ff2c3f73" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.320778 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl79r" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.322576 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b919-account-create-pjsgg" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.322585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b919-account-create-pjsgg" event={"ID":"a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a","Type":"ContainerDied","Data":"2938ffc635916be34c21847bbbeeaa98f5e78e5ec9115016ffe579e254cc3f4e"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.322643 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2938ffc635916be34c21847bbbeeaa98f5e78e5ec9115016ffe579e254cc3f4e" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.324907 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xsvt9" event={"ID":"b9609549-6956-4a00-a921-5ab8f6b0511f","Type":"ContainerDied","Data":"f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76"} Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.324941 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f77db9bfa8ea84501eccad5cc869e423a6e3f25d00fbcfa5831b37a386c3bc76" Nov 24 00:44:53 crc kubenswrapper[4888]: I1124 00:44:53.324947 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xsvt9" Nov 24 00:44:55 crc kubenswrapper[4888]: I1124 00:44:55.359503 4888 generic.go:334] "Generic (PLEG): container finished" podID="7ba9e0fb-846a-49dc-bdc3-a0e201324dab" containerID="0a786c9ad3f7c81d1bc703c3f11316e1f170b7a6174c40ed7f8d06f881de4a88" exitCode=0 Nov 24 00:44:55 crc kubenswrapper[4888]: I1124 00:44:55.359645 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerDied","Data":"0a786c9ad3f7c81d1bc703c3f11316e1f170b7a6174c40ed7f8d06f881de4a88"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.109225 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72247" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.158576 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8khk\" (UniqueName: \"kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk\") pod \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.158735 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data\") pod \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.158759 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle\") pod \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.158906 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data\") pod \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\" (UID: \"4ff49e8b-d3e9-453a-8adb-858fba7c7418\") " Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.169092 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk" (OuterVolumeSpecName: "kube-api-access-l8khk") pod "4ff49e8b-d3e9-453a-8adb-858fba7c7418" (UID: "4ff49e8b-d3e9-453a-8adb-858fba7c7418"). InnerVolumeSpecName "kube-api-access-l8khk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.169980 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4ff49e8b-d3e9-453a-8adb-858fba7c7418" (UID: "4ff49e8b-d3e9-453a-8adb-858fba7c7418"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.238067 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ff49e8b-d3e9-453a-8adb-858fba7c7418" (UID: "4ff49e8b-d3e9-453a-8adb-858fba7c7418"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.248339 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data" (OuterVolumeSpecName: "config-data") pod "4ff49e8b-d3e9-453a-8adb-858fba7c7418" (UID: "4ff49e8b-d3e9-453a-8adb-858fba7c7418"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.263602 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.263668 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.263690 4888 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff49e8b-d3e9-453a-8adb-858fba7c7418-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.263708 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8khk\" (UniqueName: \"kubernetes.io/projected/4ff49e8b-d3e9-453a-8adb-858fba7c7418-kube-api-access-l8khk\") on node \"crc\" DevicePath \"\"" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.371558 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerStarted","Data":"19abb6b417f882a2ef48b955a0860986621a6c423cafca31876a743f6e7dd459"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.373740 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xqt9t" event={"ID":"c1cd14a9-9f24-436f-b96f-ea9457548844","Type":"ContainerStarted","Data":"12069f009ddaee72158ce58e2208ca1dd5c0ed504c9d6de9b93c94c8fdefda1b"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.376584 4888 generic.go:334] "Generic (PLEG): container finished" podID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerID="60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046" exitCode=0 Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.376659 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" event={"ID":"82b55ab5-b40e-482d-8bae-956a94d977ec","Type":"ContainerDied","Data":"60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.376689 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" event={"ID":"82b55ab5-b40e-482d-8bae-956a94d977ec","Type":"ContainerStarted","Data":"bac4abbe40a9ef53a3724ac1885bdf866e6b9561ba9674f10376f2b8d747a46d"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.381379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72247" event={"ID":"4ff49e8b-d3e9-453a-8adb-858fba7c7418","Type":"ContainerDied","Data":"b40bff3c24f6637ea71b9533fe17c85b3e96d621d40eadb243352fcc0c79657b"} Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.381400 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b40bff3c24f6637ea71b9533fe17c85b3e96d621d40eadb243352fcc0c79657b" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.381433 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72247" Nov 24 00:44:56 crc kubenswrapper[4888]: I1124 00:44:56.397619 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xqt9t" podStartSLOduration=2.665310078 podStartE2EDuration="8.397601639s" podCreationTimestamp="2025-11-24 00:44:48 +0000 UTC" firstStartedPulling="2025-11-24 00:44:50.246400093 +0000 UTC m=+1192.829084137" lastFinishedPulling="2025-11-24 00:44:55.978691634 +0000 UTC m=+1198.561375698" observedRunningTime="2025-11-24 00:44:56.389225665 +0000 UTC m=+1198.971909719" watchObservedRunningTime="2025-11-24 00:44:56.397601639 +0000 UTC m=+1198.980285673" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.396442 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" event={"ID":"82b55ab5-b40e-482d-8bae-956a94d977ec","Type":"ContainerStarted","Data":"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a"} Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.396643 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.734016 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" podStartSLOduration=6.734001103 podStartE2EDuration="6.734001103s" podCreationTimestamp="2025-11-24 00:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:44:57.428379487 +0000 UTC m=+1200.011063581" watchObservedRunningTime="2025-11-24 00:44:57.734001103 +0000 UTC m=+1200.316685147" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.734483 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812185 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812567 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61345e71-eb59-4bef-a890-65ed5a2e99b7" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812579 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="61345e71-eb59-4bef-a890-65ed5a2e99b7" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812593 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05103e58-6551-4742-ba88-12e3669fbee3" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812598 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="05103e58-6551-4742-ba88-12e3669fbee3" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812625 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff49e8b-d3e9-453a-8adb-858fba7c7418" containerName="glance-db-sync" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812632 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff49e8b-d3e9-453a-8adb-858fba7c7418" containerName="glance-db-sync" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812648 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9609549-6956-4a00-a921-5ab8f6b0511f" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812654 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9609549-6956-4a00-a921-5ab8f6b0511f" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812668 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812674 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812680 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812686 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812695 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d1470a-a999-4afe-a7e4-514914054767" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812701 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d1470a-a999-4afe-a7e4-514914054767" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812707 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3294b25b-9e52-43e0-ba6e-87b510640945" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812713 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3294b25b-9e52-43e0-ba6e-87b510640945" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: E1124 00:44:57.812726 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812731 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812917 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812932 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="61345e71-eb59-4bef-a890-65ed5a2e99b7" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812943 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d1470a-a999-4afe-a7e4-514914054767" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812951 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9609549-6956-4a00-a921-5ab8f6b0511f" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812963 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812975 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" containerName="mariadb-account-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812983 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3294b25b-9e52-43e0-ba6e-87b510640945" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.812994 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="05103e58-6551-4742-ba88-12e3669fbee3" containerName="mariadb-database-create" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.813000 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff49e8b-d3e9-453a-8adb-858fba7c7418" containerName="glance-db-sync" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.813915 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.844866 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.902956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.903254 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.903392 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.903484 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.903566 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:57 crc kubenswrapper[4888]: I1124 00:44:57.903647 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bmm\" (UniqueName: \"kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005098 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005160 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005189 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bmm\" (UniqueName: \"kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005335 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005372 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.005429 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.006539 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.006660 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.006671 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.006878 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.006969 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.029569 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bmm\" (UniqueName: \"kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm\") pod \"dnsmasq-dns-74f6bcbc87-f7k59\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.169635 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:44:58 crc kubenswrapper[4888]: I1124 00:44:58.718077 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:44:58 crc kubenswrapper[4888]: W1124 00:44:58.724839 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37fddbda_57f6_4fce_9193_1a201d220c51.slice/crio-119c67cc4bf940763336afdcdc983eaab6e5173b142b4e86662b47fdf9015adf WatchSource:0}: Error finding container 119c67cc4bf940763336afdcdc983eaab6e5173b142b4e86662b47fdf9015adf: Status 404 returned error can't find the container with id 119c67cc4bf940763336afdcdc983eaab6e5173b142b4e86662b47fdf9015adf Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.423626 4888 generic.go:334] "Generic (PLEG): container finished" podID="37fddbda-57f6-4fce-9193-1a201d220c51" containerID="fefe5644c7e356e9f49a057c446420887954f309db7a82bb78d2e596329344f9" exitCode=0 Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.424436 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" event={"ID":"37fddbda-57f6-4fce-9193-1a201d220c51","Type":"ContainerDied","Data":"fefe5644c7e356e9f49a057c446420887954f309db7a82bb78d2e596329344f9"} Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.424486 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" event={"ID":"37fddbda-57f6-4fce-9193-1a201d220c51","Type":"ContainerStarted","Data":"119c67cc4bf940763336afdcdc983eaab6e5173b142b4e86662b47fdf9015adf"} Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.428627 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerStarted","Data":"182df81fabd03e9fd27e221627260e87f2340eae084a35f11fa79eaf6c542dec"} Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.428775 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="dnsmasq-dns" containerID="cri-o://131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a" gracePeriod=10 Nov 24 00:44:59 crc kubenswrapper[4888]: I1124 00:44:59.932781 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.039348 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.040524 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.040616 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.040708 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.040772 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.040851 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-229wc\" (UniqueName: \"kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc\") pod \"82b55ab5-b40e-482d-8bae-956a94d977ec\" (UID: \"82b55ab5-b40e-482d-8bae-956a94d977ec\") " Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.045711 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc" (OuterVolumeSpecName: "kube-api-access-229wc") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "kube-api-access-229wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.089630 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config" (OuterVolumeSpecName: "config") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.089663 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.093414 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.095944 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.104137 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82b55ab5-b40e-482d-8bae-956a94d977ec" (UID: "82b55ab5-b40e-482d-8bae-956a94d977ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.131966 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47"] Nov 24 00:45:00 crc kubenswrapper[4888]: E1124 00:45:00.132343 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="init" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.132357 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="init" Nov 24 00:45:00 crc kubenswrapper[4888]: E1124 00:45:00.132366 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="dnsmasq-dns" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.132373 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="dnsmasq-dns" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.132537 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerName="dnsmasq-dns" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.133126 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.135029 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.139160 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143216 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143253 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143266 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143278 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143291 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-229wc\" (UniqueName: \"kubernetes.io/projected/82b55ab5-b40e-482d-8bae-956a94d977ec-kube-api-access-229wc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.143301 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82b55ab5-b40e-482d-8bae-956a94d977ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.159093 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47"] Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.244427 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.244498 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.244558 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plhbb\" (UniqueName: \"kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.352174 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.352281 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.352345 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plhbb\" (UniqueName: \"kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.354163 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.358782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.394445 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plhbb\" (UniqueName: \"kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb\") pod \"collect-profiles-29399085-wzc47\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.438432 4888 generic.go:334] "Generic (PLEG): container finished" podID="c1cd14a9-9f24-436f-b96f-ea9457548844" containerID="12069f009ddaee72158ce58e2208ca1dd5c0ed504c9d6de9b93c94c8fdefda1b" exitCode=0 Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.438489 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xqt9t" event={"ID":"c1cd14a9-9f24-436f-b96f-ea9457548844","Type":"ContainerDied","Data":"12069f009ddaee72158ce58e2208ca1dd5c0ed504c9d6de9b93c94c8fdefda1b"} Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.440961 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" event={"ID":"37fddbda-57f6-4fce-9193-1a201d220c51","Type":"ContainerStarted","Data":"6ad8824a1fceb98deb884eea157fa7c2d29c59f148d717923daa999428b646ca"} Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.441653 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.443785 4888 generic.go:334] "Generic (PLEG): container finished" podID="82b55ab5-b40e-482d-8bae-956a94d977ec" containerID="131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a" exitCode=0 Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.443883 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" event={"ID":"82b55ab5-b40e-482d-8bae-956a94d977ec","Type":"ContainerDied","Data":"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a"} Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.443902 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" event={"ID":"82b55ab5-b40e-482d-8bae-956a94d977ec","Type":"ContainerDied","Data":"bac4abbe40a9ef53a3724ac1885bdf866e6b9561ba9674f10376f2b8d747a46d"} Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.443920 4888 scope.go:117] "RemoveContainer" containerID="131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.443999 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rfmwb" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.452213 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7ba9e0fb-846a-49dc-bdc3-a0e201324dab","Type":"ContainerStarted","Data":"72154cc9055195bc03a18981ff0dc0310385223276733f8a7bb8a21ffb623d2c"} Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.453184 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.525537 4888 scope.go:117] "RemoveContainer" containerID="60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.530733 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" podStartSLOduration=3.530710758 podStartE2EDuration="3.530710758s" podCreationTimestamp="2025-11-24 00:44:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:00.51792237 +0000 UTC m=+1203.100606414" watchObservedRunningTime="2025-11-24 00:45:00.530710758 +0000 UTC m=+1203.113394812" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.578505 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.578488294 podStartE2EDuration="17.578488294s" podCreationTimestamp="2025-11-24 00:44:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:00.573190476 +0000 UTC m=+1203.155874520" watchObservedRunningTime="2025-11-24 00:45:00.578488294 +0000 UTC m=+1203.161172338" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.612915 4888 scope.go:117] "RemoveContainer" containerID="131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.619855 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:45:00 crc kubenswrapper[4888]: E1124 00:45:00.622947 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a\": container with ID starting with 131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a not found: ID does not exist" containerID="131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.622989 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a"} err="failed to get container status \"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a\": rpc error: code = NotFound desc = could not find container \"131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a\": container with ID starting with 131c983737846f04d8722b254d33b9c91541d56d8406df46116f93e204d11b6a not found: ID does not exist" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.623015 4888 scope.go:117] "RemoveContainer" containerID="60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046" Nov 24 00:45:00 crc kubenswrapper[4888]: E1124 00:45:00.624782 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046\": container with ID starting with 60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046 not found: ID does not exist" containerID="60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.624833 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046"} err="failed to get container status \"60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046\": rpc error: code = NotFound desc = could not find container \"60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046\": container with ID starting with 60eb1db6b99b2a3ec59f7b2465d218ca45efaa1fb360165599617f7057d25046 not found: ID does not exist" Nov 24 00:45:00 crc kubenswrapper[4888]: I1124 00:45:00.625601 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rfmwb"] Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.024579 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47"] Nov 24 00:45:01 crc kubenswrapper[4888]: W1124 00:45:01.031024 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8584e6b2_f5e9_4395_a596_33ead8945cf0.slice/crio-cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506 WatchSource:0}: Error finding container cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506: Status 404 returned error can't find the container with id cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506 Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.462161 4888 generic.go:334] "Generic (PLEG): container finished" podID="8584e6b2-f5e9-4395-a596-33ead8945cf0" containerID="7662f3f310053db964245230730762648b4857c92e564ac9c10e4407e0c35db1" exitCode=0 Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.462233 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" event={"ID":"8584e6b2-f5e9-4395-a596-33ead8945cf0","Type":"ContainerDied","Data":"7662f3f310053db964245230730762648b4857c92e564ac9c10e4407e0c35db1"} Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.462265 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" event={"ID":"8584e6b2-f5e9-4395-a596-33ead8945cf0","Type":"ContainerStarted","Data":"cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506"} Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.912744 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.987785 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle\") pod \"c1cd14a9-9f24-436f-b96f-ea9457548844\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.988014 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data\") pod \"c1cd14a9-9f24-436f-b96f-ea9457548844\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.988106 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wnb8\" (UniqueName: \"kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8\") pod \"c1cd14a9-9f24-436f-b96f-ea9457548844\" (UID: \"c1cd14a9-9f24-436f-b96f-ea9457548844\") " Nov 24 00:45:01 crc kubenswrapper[4888]: I1124 00:45:01.994671 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8" (OuterVolumeSpecName: "kube-api-access-8wnb8") pod "c1cd14a9-9f24-436f-b96f-ea9457548844" (UID: "c1cd14a9-9f24-436f-b96f-ea9457548844"). InnerVolumeSpecName "kube-api-access-8wnb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.049024 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data" (OuterVolumeSpecName: "config-data") pod "c1cd14a9-9f24-436f-b96f-ea9457548844" (UID: "c1cd14a9-9f24-436f-b96f-ea9457548844"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.053564 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1cd14a9-9f24-436f-b96f-ea9457548844" (UID: "c1cd14a9-9f24-436f-b96f-ea9457548844"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.090106 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.090282 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cd14a9-9f24-436f-b96f-ea9457548844-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.090433 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wnb8\" (UniqueName: \"kubernetes.io/projected/c1cd14a9-9f24-436f-b96f-ea9457548844-kube-api-access-8wnb8\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.278643 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b55ab5-b40e-482d-8bae-956a94d977ec" path="/var/lib/kubelet/pods/82b55ab5-b40e-482d-8bae-956a94d977ec/volumes" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.475649 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xqt9t" event={"ID":"c1cd14a9-9f24-436f-b96f-ea9457548844","Type":"ContainerDied","Data":"7ac004f2d8f253c763350e89b6a8b95a6d10689f1ddd503ef31926a7eb0a4ff5"} Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.475709 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ac004f2d8f253c763350e89b6a8b95a6d10689f1ddd503ef31926a7eb0a4ff5" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.475681 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xqt9t" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.745100 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.745324 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="dnsmasq-dns" containerID="cri-o://6ad8824a1fceb98deb884eea157fa7c2d29c59f148d717923daa999428b646ca" gracePeriod=10 Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.764664 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4wtsz"] Nov 24 00:45:02 crc kubenswrapper[4888]: E1124 00:45:02.765204 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cd14a9-9f24-436f-b96f-ea9457548844" containerName="keystone-db-sync" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.765223 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cd14a9-9f24-436f-b96f-ea9457548844" containerName="keystone-db-sync" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.765445 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cd14a9-9f24-436f-b96f-ea9457548844" containerName="keystone-db-sync" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.766543 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.771386 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.771669 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.771806 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-pjwx7" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.772002 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.772681 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.807949 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.808060 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ww44\" (UniqueName: \"kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.808097 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.808128 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.808179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.808231 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.852603 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4wtsz"] Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.912483 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.912792 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.924713 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.929643 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.929860 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ww44\" (UniqueName: \"kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.929911 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.937917 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.947690 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.950608 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:02 crc kubenswrapper[4888]: I1124 00:45:02.959278 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.034362 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.034530 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ww44\" (UniqueName: \"kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.035181 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys\") pod \"keystone-bootstrap-4wtsz\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.058293 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.105036 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.142879 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-2qlfb"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.144143 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.149520 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.166083 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2qlfb"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.180276 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.180466 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tt2jh" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.192262 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.214917 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-pnrn8"] Nov 24 00:45:03 crc kubenswrapper[4888]: E1124 00:45:03.215316 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8584e6b2-f5e9-4395-a596-33ead8945cf0" containerName="collect-profiles" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.215331 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8584e6b2-f5e9-4395-a596-33ead8945cf0" containerName="collect-profiles" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.215621 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8584e6b2-f5e9-4395-a596-33ead8945cf0" containerName="collect-profiles" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.216291 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.221247 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.221473 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.226888 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-pnrn8"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.233070 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ghh5w" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250152 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250196 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250219 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9452\" (UniqueName: \"kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250238 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhr9b\" (UniqueName: \"kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250263 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250284 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250308 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250342 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.250418 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.273640 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jzk9c"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.277572 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.280936 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.281185 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.286556 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w7kcx" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.298931 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jzk9c"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.355065 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plhbb\" (UniqueName: \"kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb\") pod \"8584e6b2-f5e9-4395-a596-33ead8945cf0\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.355896 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume\") pod \"8584e6b2-f5e9-4395-a596-33ead8945cf0\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.355925 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume\") pod \"8584e6b2-f5e9-4395-a596-33ead8945cf0\" (UID: \"8584e6b2-f5e9-4395-a596-33ead8945cf0\") " Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356047 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356096 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356121 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356143 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356172 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356187 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9452\" (UniqueName: \"kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356206 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhr9b\" (UniqueName: \"kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356226 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356243 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356259 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356278 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356299 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75nbw\" (UniqueName: \"kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356326 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356350 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356461 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2kg9\" (UniqueName: \"kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356480 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.356505 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.357718 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.359293 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-w7rnb"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.360543 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.364329 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.364346 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2lj5b" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.365871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.366109 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.366328 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.366995 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.375242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume" (OuterVolumeSpecName: "config-volume") pod "8584e6b2-f5e9-4395-a596-33ead8945cf0" (UID: "8584e6b2-f5e9-4395-a596-33ead8945cf0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.377692 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.378416 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.395132 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8584e6b2-f5e9-4395-a596-33ead8945cf0" (UID: "8584e6b2-f5e9-4395-a596-33ead8945cf0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.397547 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb" (OuterVolumeSpecName: "kube-api-access-plhbb") pod "8584e6b2-f5e9-4395-a596-33ead8945cf0" (UID: "8584e6b2-f5e9-4395-a596-33ead8945cf0"). InnerVolumeSpecName "kube-api-access-plhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.397617 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhr9b\" (UniqueName: \"kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b\") pod \"dnsmasq-dns-847c4cc679-5gqsp\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.395877 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-w7rnb"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.424098 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-pkgvs"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.433852 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.442998 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.443202 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xck2c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.443325 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.446345 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9452\" (UniqueName: \"kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452\") pod \"heat-db-sync-2qlfb\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.453881 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pkgvs"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457799 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457859 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457898 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457926 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75nbw\" (UniqueName: \"kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457955 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.457977 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458022 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2kg9\" (UniqueName: \"kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458040 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458082 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458131 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plhbb\" (UniqueName: \"kubernetes.io/projected/8584e6b2-f5e9-4395-a596-33ead8945cf0-kube-api-access-plhbb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458141 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8584e6b2-f5e9-4395-a596-33ead8945cf0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.458150 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8584e6b2-f5e9-4395-a596-33ead8945cf0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.463773 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.472646 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.475522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.477187 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.477666 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.481495 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.482392 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.482459 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.497522 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.517371 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2kg9\" (UniqueName: \"kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9\") pod \"cinder-db-sync-jzk9c\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.518213 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75nbw\" (UniqueName: \"kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw\") pod \"neutron-db-sync-pnrn8\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.546759 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" event={"ID":"8584e6b2-f5e9-4395-a596-33ead8945cf0","Type":"ContainerDied","Data":"cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506"} Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.546826 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad0680116331935521a8a52d00a4c09151008388e4e9aa218119cf117041506" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.546909 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560097 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560200 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560222 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560256 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbvqc\" (UniqueName: \"kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560299 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560332 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.560361 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvbc5\" (UniqueName: \"kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.574447 4888 generic.go:334] "Generic (PLEG): container finished" podID="37fddbda-57f6-4fce-9193-1a201d220c51" containerID="6ad8824a1fceb98deb884eea157fa7c2d29c59f148d717923daa999428b646ca" exitCode=0 Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.574482 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" event={"ID":"37fddbda-57f6-4fce-9193-1a201d220c51","Type":"ContainerDied","Data":"6ad8824a1fceb98deb884eea157fa7c2d29c59f148d717923daa999428b646ca"} Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.596422 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.599574 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.615727 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.639869 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.649362 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.663316 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.664878 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvbc5\" (UniqueName: \"kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.664931 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.664999 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.665021 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.665043 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.665079 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbvqc\" (UniqueName: \"kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.665125 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.665163 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.666108 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.675479 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.677523 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.677877 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.680002 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.682584 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.682885 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.684135 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.688164 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbvqc\" (UniqueName: \"kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc\") pod \"placement-db-sync-pkgvs\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.689192 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvbc5\" (UniqueName: \"kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5\") pod \"barbican-db-sync-w7rnb\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.696550 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.699622 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.699874 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.715998 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.758918 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768317 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768374 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768402 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768471 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkr44\" (UniqueName: \"kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768502 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.768549 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.785925 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870670 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870725 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870803 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmff\" (UniqueName: \"kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870836 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870855 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870873 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870894 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870919 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870944 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkr44\" (UniqueName: \"kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870964 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.870984 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.871022 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.871037 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.872766 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.873461 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.874333 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.875127 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.875629 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.891040 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkr44\" (UniqueName: \"kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44\") pod \"dnsmasq-dns-785d8bcb8c-2v4q8\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973309 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flmff\" (UniqueName: \"kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973351 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973385 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973421 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973465 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973505 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973526 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.973918 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.977384 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.978606 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.991752 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.991752 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.993611 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.995005 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:03 crc kubenswrapper[4888]: I1124 00:45:03.996374 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flmff\" (UniqueName: \"kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff\") pod \"ceilometer-0\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " pod="openstack/ceilometer-0" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.062836 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.100984 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4wtsz"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.145129 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.326841 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.431410 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-w7rnb"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.483710 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.483799 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.484355 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.484412 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7bmm\" (UniqueName: \"kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.484514 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.484563 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb\") pod \"37fddbda-57f6-4fce-9193-1a201d220c51\" (UID: \"37fddbda-57f6-4fce-9193-1a201d220c51\") " Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.538575 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm" (OuterVolumeSpecName: "kube-api-access-t7bmm") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "kube-api-access-t7bmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.589786 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7bmm\" (UniqueName: \"kubernetes.io/projected/37fddbda-57f6-4fce-9193-1a201d220c51-kube-api-access-t7bmm\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.623800 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.640244 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.653252 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" event={"ID":"9d273792-fa72-4430-a0cd-bfbe8752026a","Type":"ContainerStarted","Data":"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.653313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" event={"ID":"9d273792-fa72-4430-a0cd-bfbe8752026a","Type":"ContainerStarted","Data":"e30ede15d2206cac63506f34836afaa422e68c482bfd8848403d73fecddff6ce"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.658942 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.664191 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4wtsz" event={"ID":"de3b500c-b7d8-4274-962c-aef390cc3ec5","Type":"ContainerStarted","Data":"0a660a955587e6605222ffbc6accc58c21bc2de4195fe4909a40f8bb2e9fef58"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.664296 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4wtsz" event={"ID":"de3b500c-b7d8-4274-962c-aef390cc3ec5","Type":"ContainerStarted","Data":"e29c8055fab791608ee13b2266be1acb69e0608d12aa04f7bfd820b5400111de"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.665146 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-pnrn8"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.671285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w7rnb" event={"ID":"a76821af-c9e6-47a8-a167-1531770cc394","Type":"ContainerStarted","Data":"826d9cada0d1cbadb178881371d9f861e77b8699ab1b01b31692fb208a6aa62a"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.673400 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config" (OuterVolumeSpecName: "config") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.676601 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37fddbda-57f6-4fce-9193-1a201d220c51" (UID: "37fddbda-57f6-4fce-9193-1a201d220c51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.678175 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" event={"ID":"37fddbda-57f6-4fce-9193-1a201d220c51","Type":"ContainerDied","Data":"119c67cc4bf940763336afdcdc983eaab6e5173b142b4e86662b47fdf9015adf"} Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.678244 4888 scope.go:117] "RemoveContainer" containerID="6ad8824a1fceb98deb884eea157fa7c2d29c59f148d717923daa999428b646ca" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.678427 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-f7k59" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.679830 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2qlfb"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.695880 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.696498 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.696565 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.696620 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.696673 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37fddbda-57f6-4fce-9193-1a201d220c51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.745996 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4wtsz" podStartSLOduration=2.745971002 podStartE2EDuration="2.745971002s" podCreationTimestamp="2025-11-24 00:45:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:04.73803815 +0000 UTC m=+1207.320722214" watchObservedRunningTime="2025-11-24 00:45:04.745971002 +0000 UTC m=+1207.328655046" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.769275 4888 scope.go:117] "RemoveContainer" containerID="fefe5644c7e356e9f49a057c446420887954f309db7a82bb78d2e596329344f9" Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.799760 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.825183 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-f7k59"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.866418 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jzk9c"] Nov 24 00:45:04 crc kubenswrapper[4888]: I1124 00:45:04.899212 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pkgvs"] Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.000764 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.011312 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.248761 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.317941 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.318110 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhr9b\" (UniqueName: \"kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.318901 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.318981 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.319052 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.319346 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb\") pod \"9d273792-fa72-4430-a0cd-bfbe8752026a\" (UID: \"9d273792-fa72-4430-a0cd-bfbe8752026a\") " Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.324307 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b" (OuterVolumeSpecName: "kube-api-access-bhr9b") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "kube-api-access-bhr9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.427466 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhr9b\" (UniqueName: \"kubernetes.io/projected/9d273792-fa72-4430-a0cd-bfbe8752026a-kube-api-access-bhr9b\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.429177 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.453310 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.464292 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config" (OuterVolumeSpecName: "config") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.465276 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.465454 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9d273792-fa72-4430-a0cd-bfbe8752026a" (UID: "9d273792-fa72-4430-a0cd-bfbe8752026a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.529208 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.529499 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.529509 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.529517 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.529527 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d273792-fa72-4430-a0cd-bfbe8752026a-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.702651 4888 generic.go:334] "Generic (PLEG): container finished" podID="9d273792-fa72-4430-a0cd-bfbe8752026a" containerID="8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62" exitCode=0 Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.702716 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" event={"ID":"9d273792-fa72-4430-a0cd-bfbe8752026a","Type":"ContainerDied","Data":"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.702759 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" event={"ID":"9d273792-fa72-4430-a0cd-bfbe8752026a","Type":"ContainerDied","Data":"e30ede15d2206cac63506f34836afaa422e68c482bfd8848403d73fecddff6ce"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.702777 4888 scope.go:117] "RemoveContainer" containerID="8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.702905 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-5gqsp" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.705156 4888 generic.go:334] "Generic (PLEG): container finished" podID="43659f95-de1a-41a7-b502-fd814dbac64a" containerID="af2948780bd3f2b73a8921acdf3979ecc822211ca95d831eb7678f81d02a4df9" exitCode=0 Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.705207 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" event={"ID":"43659f95-de1a-41a7-b502-fd814dbac64a","Type":"ContainerDied","Data":"af2948780bd3f2b73a8921acdf3979ecc822211ca95d831eb7678f81d02a4df9"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.705233 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" event={"ID":"43659f95-de1a-41a7-b502-fd814dbac64a","Type":"ContainerStarted","Data":"5a454c2f87eb69d7e3e980e152e27bcd42796e693afa07e8630a6b972f797a10"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.714721 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2qlfb" event={"ID":"f5cc9373-09e2-4cae-82e0-badb0da1e0ad","Type":"ContainerStarted","Data":"15d73d15e5de8afae3a6ccff999c62eb270814b5e65db6df220be7c8e94a30c3"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.716577 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerStarted","Data":"46ca956f9dcd6d925d2b256de7efbb1c8fc287685c50cd69d5b6e7b205bc8112"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.718086 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkgvs" event={"ID":"e87929ab-dd6b-408f-80ed-e1716b5384dc","Type":"ContainerStarted","Data":"1817d7c843e5adcbf959a7b9fd9efdd59592de8408e5008fbe36e01095ddc6da"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.720600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jzk9c" event={"ID":"1d5a8488-3848-4285-96bc-6313cc426ad5","Type":"ContainerStarted","Data":"8bb9989d005364700bba60b490d3abb482a4cfa0d7dea27c87dbb267a33ad9f2"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.724946 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pnrn8" event={"ID":"922fc13b-8e48-4729-bf76-2244367cdce4","Type":"ContainerStarted","Data":"b97431cb0245a05e90ed1beeae629ae67ee41da52d41e9dd0163f46e7b1ecaff"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.724985 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pnrn8" event={"ID":"922fc13b-8e48-4729-bf76-2244367cdce4","Type":"ContainerStarted","Data":"7ec872cd00a912eee4ba6ba2afb1103f32f7a8b45ac974f12b1746b59f0e44b4"} Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.756448 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-pnrn8" podStartSLOduration=2.756425771 podStartE2EDuration="2.756425771s" podCreationTimestamp="2025-11-24 00:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:05.750236558 +0000 UTC m=+1208.332920602" watchObservedRunningTime="2025-11-24 00:45:05.756425771 +0000 UTC m=+1208.339109815" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.768601 4888 scope.go:117] "RemoveContainer" containerID="8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62" Nov 24 00:45:05 crc kubenswrapper[4888]: E1124 00:45:05.775027 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62\": container with ID starting with 8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62 not found: ID does not exist" containerID="8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.775069 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62"} err="failed to get container status \"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62\": rpc error: code = NotFound desc = could not find container \"8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62\": container with ID starting with 8446511c3f491fa368ea824326ef066e1c0314b45b308c3ab9378d3f8d312a62 not found: ID does not exist" Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.795572 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:05 crc kubenswrapper[4888]: I1124 00:45:05.803428 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-5gqsp"] Nov 24 00:45:06 crc kubenswrapper[4888]: I1124 00:45:06.273293 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" path="/var/lib/kubelet/pods/37fddbda-57f6-4fce-9193-1a201d220c51/volumes" Nov 24 00:45:06 crc kubenswrapper[4888]: I1124 00:45:06.274636 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d273792-fa72-4430-a0cd-bfbe8752026a" path="/var/lib/kubelet/pods/9d273792-fa72-4430-a0cd-bfbe8752026a/volumes" Nov 24 00:45:06 crc kubenswrapper[4888]: I1124 00:45:06.759205 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" event={"ID":"43659f95-de1a-41a7-b502-fd814dbac64a","Type":"ContainerStarted","Data":"e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac"} Nov 24 00:45:06 crc kubenswrapper[4888]: I1124 00:45:06.760888 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:06 crc kubenswrapper[4888]: I1124 00:45:06.794902 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" podStartSLOduration=3.794884723 podStartE2EDuration="3.794884723s" podCreationTimestamp="2025-11-24 00:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:06.781056476 +0000 UTC m=+1209.363740530" watchObservedRunningTime="2025-11-24 00:45:06.794884723 +0000 UTC m=+1209.377568767" Nov 24 00:45:07 crc kubenswrapper[4888]: I1124 00:45:07.449120 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:08 crc kubenswrapper[4888]: I1124 00:45:08.787484 4888 generic.go:334] "Generic (PLEG): container finished" podID="de3b500c-b7d8-4274-962c-aef390cc3ec5" containerID="0a660a955587e6605222ffbc6accc58c21bc2de4195fe4909a40f8bb2e9fef58" exitCode=0 Nov 24 00:45:08 crc kubenswrapper[4888]: I1124 00:45:08.787533 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4wtsz" event={"ID":"de3b500c-b7d8-4274-962c-aef390cc3ec5","Type":"ContainerDied","Data":"0a660a955587e6605222ffbc6accc58c21bc2de4195fe4909a40f8bb2e9fef58"} Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.378946 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440064 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440154 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440213 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ww44\" (UniqueName: \"kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440331 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440357 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.440663 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts\") pod \"de3b500c-b7d8-4274-962c-aef390cc3ec5\" (UID: \"de3b500c-b7d8-4274-962c-aef390cc3ec5\") " Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.450295 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.450438 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.450686 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44" (OuterVolumeSpecName: "kube-api-access-6ww44") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "kube-api-access-6ww44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.457190 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts" (OuterVolumeSpecName: "scripts") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.489069 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data" (OuterVolumeSpecName: "config-data") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.492248 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de3b500c-b7d8-4274-962c-aef390cc3ec5" (UID: "de3b500c-b7d8-4274-962c-aef390cc3ec5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543769 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543825 4888 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543841 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543850 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ww44\" (UniqueName: \"kubernetes.io/projected/de3b500c-b7d8-4274-962c-aef390cc3ec5-kube-api-access-6ww44\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543860 4888 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.543868 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b500c-b7d8-4274-962c-aef390cc3ec5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.817874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4wtsz" event={"ID":"de3b500c-b7d8-4274-962c-aef390cc3ec5","Type":"ContainerDied","Data":"e29c8055fab791608ee13b2266be1acb69e0608d12aa04f7bfd820b5400111de"} Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.817915 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e29c8055fab791608ee13b2266be1acb69e0608d12aa04f7bfd820b5400111de" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.817967 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4wtsz" Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.888940 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4wtsz"] Nov 24 00:45:10 crc kubenswrapper[4888]: I1124 00:45:10.896733 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4wtsz"] Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006179 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nh84n"] Nov 24 00:45:11 crc kubenswrapper[4888]: E1124 00:45:11.006549 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d273792-fa72-4430-a0cd-bfbe8752026a" containerName="init" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006559 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d273792-fa72-4430-a0cd-bfbe8752026a" containerName="init" Nov 24 00:45:11 crc kubenswrapper[4888]: E1124 00:45:11.006573 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="dnsmasq-dns" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006580 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="dnsmasq-dns" Nov 24 00:45:11 crc kubenswrapper[4888]: E1124 00:45:11.006607 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="init" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006613 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="init" Nov 24 00:45:11 crc kubenswrapper[4888]: E1124 00:45:11.006634 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3b500c-b7d8-4274-962c-aef390cc3ec5" containerName="keystone-bootstrap" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006641 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3b500c-b7d8-4274-962c-aef390cc3ec5" containerName="keystone-bootstrap" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006819 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="37fddbda-57f6-4fce-9193-1a201d220c51" containerName="dnsmasq-dns" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006833 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d273792-fa72-4430-a0cd-bfbe8752026a" containerName="init" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.006850 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3b500c-b7d8-4274-962c-aef390cc3ec5" containerName="keystone-bootstrap" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.007530 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.036749 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nh84n"] Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.051804 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.052201 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.052370 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.053758 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-pjwx7" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.054948 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.161864 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.162007 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.162034 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dk6t\" (UniqueName: \"kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.162078 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.162100 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.162153 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.263554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.263944 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.263975 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dk6t\" (UniqueName: \"kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.264002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.264025 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.264065 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.271878 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.273034 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.274798 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.280544 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.281184 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.290196 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dk6t\" (UniqueName: \"kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t\") pod \"keystone-bootstrap-nh84n\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:11 crc kubenswrapper[4888]: I1124 00:45:11.389312 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:12 crc kubenswrapper[4888]: I1124 00:45:12.255599 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3b500c-b7d8-4274-962c-aef390cc3ec5" path="/var/lib/kubelet/pods/de3b500c-b7d8-4274-962c-aef390cc3ec5/volumes" Nov 24 00:45:13 crc kubenswrapper[4888]: I1124 00:45:13.759559 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 00:45:13 crc kubenswrapper[4888]: I1124 00:45:13.771161 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 00:45:13 crc kubenswrapper[4888]: I1124 00:45:13.869843 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 00:45:13 crc kubenswrapper[4888]: I1124 00:45:13.980412 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:14 crc kubenswrapper[4888]: I1124 00:45:14.049784 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:45:14 crc kubenswrapper[4888]: I1124 00:45:14.050307 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-m5bkb" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" containerID="cri-o://e8820fc33a81d4ead6e9168c9ef6d26c2debe7de010bd4fac2a2310daba66c49" gracePeriod=10 Nov 24 00:45:14 crc kubenswrapper[4888]: I1124 00:45:14.877217 4888 generic.go:334] "Generic (PLEG): container finished" podID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerID="e8820fc33a81d4ead6e9168c9ef6d26c2debe7de010bd4fac2a2310daba66c49" exitCode=0 Nov 24 00:45:14 crc kubenswrapper[4888]: I1124 00:45:14.877282 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m5bkb" event={"ID":"cc8a2262-2ebb-4ced-bec2-2760ceb19595","Type":"ContainerDied","Data":"e8820fc33a81d4ead6e9168c9ef6d26c2debe7de010bd4fac2a2310daba66c49"} Nov 24 00:45:18 crc kubenswrapper[4888]: I1124 00:45:18.491216 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-m5bkb" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.144:5353: connect: connection refused" Nov 24 00:45:21 crc kubenswrapper[4888]: I1124 00:45:21.966859 4888 generic.go:334] "Generic (PLEG): container finished" podID="922fc13b-8e48-4729-bf76-2244367cdce4" containerID="b97431cb0245a05e90ed1beeae629ae67ee41da52d41e9dd0163f46e7b1ecaff" exitCode=0 Nov 24 00:45:21 crc kubenswrapper[4888]: I1124 00:45:21.966915 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pnrn8" event={"ID":"922fc13b-8e48-4729-bf76-2244367cdce4","Type":"ContainerDied","Data":"b97431cb0245a05e90ed1beeae629ae67ee41da52d41e9dd0163f46e7b1ecaff"} Nov 24 00:45:23 crc kubenswrapper[4888]: I1124 00:45:23.489950 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-m5bkb" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.144:5353: connect: connection refused" Nov 24 00:45:23 crc kubenswrapper[4888]: I1124 00:45:23.591510 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:45:23 crc kubenswrapper[4888]: I1124 00:45:23.591589 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:45:28 crc kubenswrapper[4888]: E1124 00:45:28.701680 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Nov 24 00:45:28 crc kubenswrapper[4888]: E1124 00:45:28.702421 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k9452,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-2qlfb_openstack(f5cc9373-09e2-4cae-82e0-badb0da1e0ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:45:28 crc kubenswrapper[4888]: E1124 00:45:28.704468 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-2qlfb" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.789521 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.794460 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.832277 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgphw\" (UniqueName: \"kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw\") pod \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.832366 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb\") pod \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.832437 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config\") pod \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.832474 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb\") pod \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.832533 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc\") pod \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\" (UID: \"cc8a2262-2ebb-4ced-bec2-2760ceb19595\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.845242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw" (OuterVolumeSpecName: "kube-api-access-xgphw") pod "cc8a2262-2ebb-4ced-bec2-2760ceb19595" (UID: "cc8a2262-2ebb-4ced-bec2-2760ceb19595"). InnerVolumeSpecName "kube-api-access-xgphw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.891052 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc8a2262-2ebb-4ced-bec2-2760ceb19595" (UID: "cc8a2262-2ebb-4ced-bec2-2760ceb19595"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.901113 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc8a2262-2ebb-4ced-bec2-2760ceb19595" (UID: "cc8a2262-2ebb-4ced-bec2-2760ceb19595"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.911009 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config" (OuterVolumeSpecName: "config") pod "cc8a2262-2ebb-4ced-bec2-2760ceb19595" (UID: "cc8a2262-2ebb-4ced-bec2-2760ceb19595"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.912623 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc8a2262-2ebb-4ced-bec2-2760ceb19595" (UID: "cc8a2262-2ebb-4ced-bec2-2760ceb19595"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.934035 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle\") pod \"922fc13b-8e48-4729-bf76-2244367cdce4\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.934317 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75nbw\" (UniqueName: \"kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw\") pod \"922fc13b-8e48-4729-bf76-2244367cdce4\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.934392 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config\") pod \"922fc13b-8e48-4729-bf76-2244367cdce4\" (UID: \"922fc13b-8e48-4729-bf76-2244367cdce4\") " Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.935181 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgphw\" (UniqueName: \"kubernetes.io/projected/cc8a2262-2ebb-4ced-bec2-2760ceb19595-kube-api-access-xgphw\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.935202 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.935214 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.935225 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.935233 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8a2262-2ebb-4ced-bec2-2760ceb19595-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.943136 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw" (OuterVolumeSpecName: "kube-api-access-75nbw") pod "922fc13b-8e48-4729-bf76-2244367cdce4" (UID: "922fc13b-8e48-4729-bf76-2244367cdce4"). InnerVolumeSpecName "kube-api-access-75nbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.956001 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "922fc13b-8e48-4729-bf76-2244367cdce4" (UID: "922fc13b-8e48-4729-bf76-2244367cdce4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:28 crc kubenswrapper[4888]: I1124 00:45:28.987950 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config" (OuterVolumeSpecName: "config") pod "922fc13b-8e48-4729-bf76-2244367cdce4" (UID: "922fc13b-8e48-4729-bf76-2244367cdce4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.036735 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.036786 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75nbw\" (UniqueName: \"kubernetes.io/projected/922fc13b-8e48-4729-bf76-2244367cdce4-kube-api-access-75nbw\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.036798 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/922fc13b-8e48-4729-bf76-2244367cdce4-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.060382 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m5bkb" event={"ID":"cc8a2262-2ebb-4ced-bec2-2760ceb19595","Type":"ContainerDied","Data":"51984e4816ab56d7d3e5592327c9fb4ab644caebee80c61094f46e00ca5cb00a"} Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.061301 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m5bkb" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.061291 4888 scope.go:117] "RemoveContainer" containerID="e8820fc33a81d4ead6e9168c9ef6d26c2debe7de010bd4fac2a2310daba66c49" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.064875 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pnrn8" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.065873 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pnrn8" event={"ID":"922fc13b-8e48-4729-bf76-2244367cdce4","Type":"ContainerDied","Data":"7ec872cd00a912eee4ba6ba2afb1103f32f7a8b45ac974f12b1746b59f0e44b4"} Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.066259 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ec872cd00a912eee4ba6ba2afb1103f32f7a8b45ac974f12b1746b59f0e44b4" Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.066720 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-2qlfb" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.110463 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.116982 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m5bkb"] Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.181451 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.181642 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n79h74hf9h574hfh5bfh5c4h57dh55bh695h68ch687h66bh688h65fh7dh584h5c5h7dh54bh7ch5d8h7ch67dh9ch76h546hfh667h58dh4h654q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-flmff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ff46e22f-f4d5-414d-b152-a262ed0ab92c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.995949 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.996666 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.996693 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.996707 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922fc13b-8e48-4729-bf76-2244367cdce4" containerName="neutron-db-sync" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.996713 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="922fc13b-8e48-4729-bf76-2244367cdce4" containerName="neutron-db-sync" Nov 24 00:45:29 crc kubenswrapper[4888]: E1124 00:45:29.996747 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="init" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.996765 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="init" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.996982 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.997007 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="922fc13b-8e48-4729-bf76-2244367cdce4" containerName="neutron-db-sync" Nov 24 00:45:29 crc kubenswrapper[4888]: I1124 00:45:29.998053 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.028259 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054542 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054611 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwsgv\" (UniqueName: \"kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054647 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054670 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054746 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.054792 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.125242 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.126918 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.134175 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.134327 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.134430 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ghh5w" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.134544 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.134945 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156632 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156702 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156762 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwsgv\" (UniqueName: \"kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156822 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156843 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.156902 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.157636 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.165387 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.167712 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.168624 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.168646 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.185799 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwsgv\" (UniqueName: \"kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv\") pod \"dnsmasq-dns-55f844cf75-bpqx2\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.255378 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" path="/var/lib/kubelet/pods/cc8a2262-2ebb-4ced-bec2-2760ceb19595/volumes" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.257967 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.258055 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.258079 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.258102 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-466kt\" (UniqueName: \"kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.258301 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.315299 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.360463 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.360512 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.360537 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-466kt\" (UniqueName: \"kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.360629 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.360699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.370509 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.372025 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.377543 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.402382 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.409250 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-466kt\" (UniqueName: \"kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt\") pod \"neutron-58bdd5575d-dsrnk\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.453129 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:30 crc kubenswrapper[4888]: I1124 00:45:30.795011 4888 scope.go:117] "RemoveContainer" containerID="d62cb800841d9a442ffe7ef47e01010a34dff0cc20cc27f85cdf489cab83d217" Nov 24 00:45:30 crc kubenswrapper[4888]: E1124 00:45:30.824122 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 00:45:30 crc kubenswrapper[4888]: E1124 00:45:30.824232 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2kg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jzk9c_openstack(1d5a8488-3848-4285-96bc-6313cc426ad5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:45:30 crc kubenswrapper[4888]: E1124 00:45:30.825437 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jzk9c" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.106251 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkgvs" event={"ID":"e87929ab-dd6b-408f-80ed-e1716b5384dc","Type":"ContainerStarted","Data":"e32ba11b3e32fdbb8345ec3932956ccfb6eec9e7d8e4068e72c2a106d8bc0631"} Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.113893 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w7rnb" event={"ID":"a76821af-c9e6-47a8-a167-1531770cc394","Type":"ContainerStarted","Data":"2c8e8dbb0ade978dd2c31689522314955b9fb1618cb0844ccfb122e1e44ee354"} Nov 24 00:45:31 crc kubenswrapper[4888]: E1124 00:45:31.114368 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-jzk9c" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.129387 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-pkgvs" podStartSLOduration=2.211185163 podStartE2EDuration="28.129368751s" podCreationTimestamp="2025-11-24 00:45:03 +0000 UTC" firstStartedPulling="2025-11-24 00:45:04.861139633 +0000 UTC m=+1207.443823677" lastFinishedPulling="2025-11-24 00:45:30.779323221 +0000 UTC m=+1233.362007265" observedRunningTime="2025-11-24 00:45:31.122101957 +0000 UTC m=+1233.704786001" watchObservedRunningTime="2025-11-24 00:45:31.129368751 +0000 UTC m=+1233.712052795" Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.185709 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-w7rnb" podStartSLOduration=1.795764745 podStartE2EDuration="28.185689846s" podCreationTimestamp="2025-11-24 00:45:03 +0000 UTC" firstStartedPulling="2025-11-24 00:45:04.43300833 +0000 UTC m=+1207.015692374" lastFinishedPulling="2025-11-24 00:45:30.822933431 +0000 UTC m=+1233.405617475" observedRunningTime="2025-11-24 00:45:31.153446904 +0000 UTC m=+1233.736130948" watchObservedRunningTime="2025-11-24 00:45:31.185689846 +0000 UTC m=+1233.768373890" Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.297754 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.312606 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nh84n"] Nov 24 00:45:31 crc kubenswrapper[4888]: I1124 00:45:31.386726 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:45:31 crc kubenswrapper[4888]: W1124 00:45:31.389858 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96baab99_3e6b_4035_85ed_acaed5a4c96f.slice/crio-17d6c2e265b22641759208996e1b06b56a1158709814ad133cd9dddf69ae60d8 WatchSource:0}: Error finding container 17d6c2e265b22641759208996e1b06b56a1158709814ad133cd9dddf69ae60d8: Status 404 returned error can't find the container with id 17d6c2e265b22641759208996e1b06b56a1158709814ad133cd9dddf69ae60d8 Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.134491 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerStarted","Data":"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.135379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerStarted","Data":"17d6c2e265b22641759208996e1b06b56a1158709814ad133cd9dddf69ae60d8"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.137600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh84n" event={"ID":"a4132e22-3a19-46cf-b764-057325ee448d","Type":"ContainerStarted","Data":"b7efcef354ddd27e8794abf82259baf7138afcabf60301155aca700caafff68c"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.137706 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh84n" event={"ID":"a4132e22-3a19-46cf-b764-057325ee448d","Type":"ContainerStarted","Data":"a2459ece4cbb53c64d2ce6f836838086fff7b0539c811831c22ec1560c6609ce"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.139268 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerStarted","Data":"579a72c66538b5334ea3c608b33b8ac7e71f84d4d7adb9bdbecd081dcb6d21df"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.142074 4888 generic.go:334] "Generic (PLEG): container finished" podID="4f2220ab-7482-436b-8386-689c044405b3" containerID="7e63107fa917588d5227a8528b69ee5c1e2683db8651aaa589458a02d317ac88" exitCode=0 Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.142226 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" event={"ID":"4f2220ab-7482-436b-8386-689c044405b3","Type":"ContainerDied","Data":"7e63107fa917588d5227a8528b69ee5c1e2683db8651aaa589458a02d317ac88"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.142332 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" event={"ID":"4f2220ab-7482-436b-8386-689c044405b3","Type":"ContainerStarted","Data":"1d5d54aa3a97f9459f78504704ab297b9f40332817621c44629a170877228e2d"} Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.158946 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nh84n" podStartSLOduration=22.158922824 podStartE2EDuration="22.158922824s" podCreationTimestamp="2025-11-24 00:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:32.152140434 +0000 UTC m=+1234.734824478" watchObservedRunningTime="2025-11-24 00:45:32.158922824 +0000 UTC m=+1234.741606868" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.508482 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-686d6d8cd9-m24dl"] Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.510324 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.513280 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.513844 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.528273 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-686d6d8cd9-m24dl"] Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.633781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-public-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634092 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-httpd-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634383 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-internal-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634540 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-ovndb-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634635 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd2ds\" (UniqueName: \"kubernetes.io/projected/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-kube-api-access-sd2ds\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.634762 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-combined-ca-bundle\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.736946 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737341 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-httpd-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737377 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-internal-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737467 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-ovndb-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737501 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd2ds\" (UniqueName: \"kubernetes.io/projected/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-kube-api-access-sd2ds\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737537 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-combined-ca-bundle\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.737634 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-public-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.752258 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.762717 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-public-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.766384 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-internal-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.766592 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-combined-ca-bundle\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.768507 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-httpd-config\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.769971 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd2ds\" (UniqueName: \"kubernetes.io/projected/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-kube-api-access-sd2ds\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.787583 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16-ovndb-tls-certs\") pod \"neutron-686d6d8cd9-m24dl\" (UID: \"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16\") " pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:32 crc kubenswrapper[4888]: I1124 00:45:32.843402 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.159873 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerStarted","Data":"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721"} Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.160414 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.165038 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" event={"ID":"4f2220ab-7482-436b-8386-689c044405b3","Type":"ContainerStarted","Data":"637fe1043e79450f555a42fea83447bf5def0711407e00f87b4463a927affed5"} Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.166092 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.204313 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58bdd5575d-dsrnk" podStartSLOduration=3.204293509 podStartE2EDuration="3.204293509s" podCreationTimestamp="2025-11-24 00:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:33.183103757 +0000 UTC m=+1235.765787801" watchObservedRunningTime="2025-11-24 00:45:33.204293509 +0000 UTC m=+1235.786977553" Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.206410 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" podStartSLOduration=4.206403708 podStartE2EDuration="4.206403708s" podCreationTimestamp="2025-11-24 00:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:33.199047123 +0000 UTC m=+1235.781731167" watchObservedRunningTime="2025-11-24 00:45:33.206403708 +0000 UTC m=+1235.789087752" Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.494479 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-686d6d8cd9-m24dl"] Nov 24 00:45:33 crc kubenswrapper[4888]: I1124 00:45:33.526535 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-m5bkb" podUID="cc8a2262-2ebb-4ced-bec2-2760ceb19595" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.144:5353: i/o timeout" Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.214921 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-686d6d8cd9-m24dl" event={"ID":"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16","Type":"ContainerStarted","Data":"5e3eec11c6eded1832970bb8cc4c447969550822ce07ccf1d1c246f5a0fff7cb"} Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.215241 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-686d6d8cd9-m24dl" event={"ID":"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16","Type":"ContainerStarted","Data":"c31bcca8bd33cfac58903f3edf462565a20699fb002cc2c751bd31af7ad6fa9d"} Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.215265 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-686d6d8cd9-m24dl" event={"ID":"d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16","Type":"ContainerStarted","Data":"762b9c549586f7ac526f7cb4e80d425580e71c3a009d87233d349a3d6f463545"} Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.216372 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.244537 4888 generic.go:334] "Generic (PLEG): container finished" podID="a76821af-c9e6-47a8-a167-1531770cc394" containerID="2c8e8dbb0ade978dd2c31689522314955b9fb1618cb0844ccfb122e1e44ee354" exitCode=0 Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.247695 4888 generic.go:334] "Generic (PLEG): container finished" podID="e87929ab-dd6b-408f-80ed-e1716b5384dc" containerID="e32ba11b3e32fdbb8345ec3932956ccfb6eec9e7d8e4068e72c2a106d8bc0631" exitCode=0 Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.272023 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-686d6d8cd9-m24dl" podStartSLOduration=2.272002489 podStartE2EDuration="2.272002489s" podCreationTimestamp="2025-11-24 00:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:34.267085971 +0000 UTC m=+1236.849770015" watchObservedRunningTime="2025-11-24 00:45:34.272002489 +0000 UTC m=+1236.854686533" Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.277874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w7rnb" event={"ID":"a76821af-c9e6-47a8-a167-1531770cc394","Type":"ContainerDied","Data":"2c8e8dbb0ade978dd2c31689522314955b9fb1618cb0844ccfb122e1e44ee354"} Nov 24 00:45:34 crc kubenswrapper[4888]: I1124 00:45:34.277915 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkgvs" event={"ID":"e87929ab-dd6b-408f-80ed-e1716b5384dc","Type":"ContainerDied","Data":"e32ba11b3e32fdbb8345ec3932956ccfb6eec9e7d8e4068e72c2a106d8bc0631"} Nov 24 00:45:36 crc kubenswrapper[4888]: I1124 00:45:36.280913 4888 generic.go:334] "Generic (PLEG): container finished" podID="a4132e22-3a19-46cf-b764-057325ee448d" containerID="b7efcef354ddd27e8794abf82259baf7138afcabf60301155aca700caafff68c" exitCode=0 Nov 24 00:45:36 crc kubenswrapper[4888]: I1124 00:45:36.281182 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh84n" event={"ID":"a4132e22-3a19-46cf-b764-057325ee448d","Type":"ContainerDied","Data":"b7efcef354ddd27e8794abf82259baf7138afcabf60301155aca700caafff68c"} Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.313453 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w7rnb" event={"ID":"a76821af-c9e6-47a8-a167-1531770cc394","Type":"ContainerDied","Data":"826d9cada0d1cbadb178881371d9f861e77b8699ab1b01b31692fb208a6aa62a"} Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.313927 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="826d9cada0d1cbadb178881371d9f861e77b8699ab1b01b31692fb208a6aa62a" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.315336 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pkgvs" event={"ID":"e87929ab-dd6b-408f-80ed-e1716b5384dc","Type":"ContainerDied","Data":"1817d7c843e5adcbf959a7b9fd9efdd59592de8408e5008fbe36e01095ddc6da"} Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.315386 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1817d7c843e5adcbf959a7b9fd9efdd59592de8408e5008fbe36e01095ddc6da" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.317619 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nh84n" event={"ID":"a4132e22-3a19-46cf-b764-057325ee448d","Type":"ContainerDied","Data":"a2459ece4cbb53c64d2ce6f836838086fff7b0539c811831c22ec1560c6609ce"} Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.317657 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2459ece4cbb53c64d2ce6f836838086fff7b0539c811831c22ec1560c6609ce" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.478483 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.483804 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.490954 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.586862 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587025 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587117 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvbc5\" (UniqueName: \"kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5\") pod \"a76821af-c9e6-47a8-a167-1531770cc394\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587193 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data\") pod \"e87929ab-dd6b-408f-80ed-e1716b5384dc\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587242 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587291 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbvqc\" (UniqueName: \"kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc\") pod \"e87929ab-dd6b-408f-80ed-e1716b5384dc\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587327 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dk6t\" (UniqueName: \"kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587388 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts\") pod \"e87929ab-dd6b-408f-80ed-e1716b5384dc\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587410 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data\") pod \"a76821af-c9e6-47a8-a167-1531770cc394\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587431 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle\") pod \"e87929ab-dd6b-408f-80ed-e1716b5384dc\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587461 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle\") pod \"a76821af-c9e6-47a8-a167-1531770cc394\" (UID: \"a76821af-c9e6-47a8-a167-1531770cc394\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587548 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587584 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys\") pod \"a4132e22-3a19-46cf-b764-057325ee448d\" (UID: \"a4132e22-3a19-46cf-b764-057325ee448d\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.587609 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs\") pod \"e87929ab-dd6b-408f-80ed-e1716b5384dc\" (UID: \"e87929ab-dd6b-408f-80ed-e1716b5384dc\") " Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.588506 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs" (OuterVolumeSpecName: "logs") pod "e87929ab-dd6b-408f-80ed-e1716b5384dc" (UID: "e87929ab-dd6b-408f-80ed-e1716b5384dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.593510 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts" (OuterVolumeSpecName: "scripts") pod "e87929ab-dd6b-408f-80ed-e1716b5384dc" (UID: "e87929ab-dd6b-408f-80ed-e1716b5384dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.593552 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts" (OuterVolumeSpecName: "scripts") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.596847 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5" (OuterVolumeSpecName: "kube-api-access-xvbc5") pod "a76821af-c9e6-47a8-a167-1531770cc394" (UID: "a76821af-c9e6-47a8-a167-1531770cc394"). InnerVolumeSpecName "kube-api-access-xvbc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.603354 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a76821af-c9e6-47a8-a167-1531770cc394" (UID: "a76821af-c9e6-47a8-a167-1531770cc394"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.603391 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc" (OuterVolumeSpecName: "kube-api-access-wbvqc") pod "e87929ab-dd6b-408f-80ed-e1716b5384dc" (UID: "e87929ab-dd6b-408f-80ed-e1716b5384dc"). InnerVolumeSpecName "kube-api-access-wbvqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.610599 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.611961 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.612002 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t" (OuterVolumeSpecName: "kube-api-access-5dk6t") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "kube-api-access-5dk6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.620082 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e87929ab-dd6b-408f-80ed-e1716b5384dc" (UID: "e87929ab-dd6b-408f-80ed-e1716b5384dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.622643 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data" (OuterVolumeSpecName: "config-data") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.623691 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data" (OuterVolumeSpecName: "config-data") pod "e87929ab-dd6b-408f-80ed-e1716b5384dc" (UID: "e87929ab-dd6b-408f-80ed-e1716b5384dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.624606 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a76821af-c9e6-47a8-a167-1531770cc394" (UID: "a76821af-c9e6-47a8-a167-1531770cc394"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.642948 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4132e22-3a19-46cf-b764-057325ee448d" (UID: "a4132e22-3a19-46cf-b764-057325ee448d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690018 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvbc5\" (UniqueName: \"kubernetes.io/projected/a76821af-c9e6-47a8-a167-1531770cc394-kube-api-access-xvbc5\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690122 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690143 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690159 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbvqc\" (UniqueName: \"kubernetes.io/projected/e87929ab-dd6b-408f-80ed-e1716b5384dc-kube-api-access-wbvqc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690177 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dk6t\" (UniqueName: \"kubernetes.io/projected/a4132e22-3a19-46cf-b764-057325ee448d-kube-api-access-5dk6t\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690191 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690210 4888 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690226 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87929ab-dd6b-408f-80ed-e1716b5384dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690251 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76821af-c9e6-47a8-a167-1531770cc394-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690267 4888 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690279 4888 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690306 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e87929ab-dd6b-408f-80ed-e1716b5384dc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690322 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:39 crc kubenswrapper[4888]: I1124 00:45:39.690336 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4132e22-3a19-46cf-b764-057325ee448d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.318082 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.331302 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerStarted","Data":"e698662ba24446ccef3f0cc1312c79317909ce77f9ef9c726d7141f83ed808e9"} Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.331365 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nh84n" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.331342 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pkgvs" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.331447 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w7rnb" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.412575 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.412940 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="dnsmasq-dns" containerID="cri-o://e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac" gracePeriod=10 Nov 24 00:45:40 crc kubenswrapper[4888]: E1124 00:45:40.595221 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43659f95_de1a_41a7_b502_fd814dbac64a.slice/crio-e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.623451 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-669f454766-khh4h"] Nov 24 00:45:40 crc kubenswrapper[4888]: E1124 00:45:40.623875 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4132e22-3a19-46cf-b764-057325ee448d" containerName="keystone-bootstrap" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.623892 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4132e22-3a19-46cf-b764-057325ee448d" containerName="keystone-bootstrap" Nov 24 00:45:40 crc kubenswrapper[4888]: E1124 00:45:40.623908 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87929ab-dd6b-408f-80ed-e1716b5384dc" containerName="placement-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.623915 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87929ab-dd6b-408f-80ed-e1716b5384dc" containerName="placement-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: E1124 00:45:40.623938 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76821af-c9e6-47a8-a167-1531770cc394" containerName="barbican-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.624008 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76821af-c9e6-47a8-a167-1531770cc394" containerName="barbican-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.628240 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76821af-c9e6-47a8-a167-1531770cc394" containerName="barbican-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.628459 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4132e22-3a19-46cf-b764-057325ee448d" containerName="keystone-bootstrap" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.628544 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87929ab-dd6b-408f-80ed-e1716b5384dc" containerName="placement-db-sync" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.632143 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.649285 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.649746 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.649998 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.650836 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xck2c" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.651837 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.725130 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-669f454766-khh4h"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726416 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-scripts\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726539 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-internal-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726592 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-public-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726664 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-combined-ca-bundle\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726722 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c22d012-f519-4556-887a-2a0a11d7f3ec-logs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726896 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-config-data\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.726958 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dk8s\" (UniqueName: \"kubernetes.io/projected/4c22d012-f519-4556-887a-2a0a11d7f3ec-kube-api-access-9dk8s\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.768161 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b7f898bdd-bbfm7"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.769630 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.779796 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.780133 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.780379 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-pjwx7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.780584 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.780750 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.787100 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.803114 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b7f898bdd-bbfm7"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.832566 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-public-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.832793 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrk6z\" (UniqueName: \"kubernetes.io/projected/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-kube-api-access-rrk6z\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.832876 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-scripts\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.832956 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-config-data\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833032 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-combined-ca-bundle\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833199 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-scripts\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-internal-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833528 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-credential-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833595 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-public-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833706 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-combined-ca-bundle\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833768 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-fernet-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833832 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c22d012-f519-4556-887a-2a0a11d7f3ec-logs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.833862 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-internal-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.834107 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-config-data\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.834190 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dk8s\" (UniqueName: \"kubernetes.io/projected/4c22d012-f519-4556-887a-2a0a11d7f3ec-kube-api-access-9dk8s\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.842183 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-public-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.845620 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c22d012-f519-4556-887a-2a0a11d7f3ec-logs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.846009 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-scripts\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.847955 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-combined-ca-bundle\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.865825 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-config-data\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.874634 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c22d012-f519-4556-887a-2a0a11d7f3ec-internal-tls-certs\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.884575 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-96d8b7cb6-6rvkq"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.898029 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.939736 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940062 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2lj5b" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940181 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940897 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-public-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrk6z\" (UniqueName: \"kubernetes.io/projected/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-kube-api-access-rrk6z\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940959 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-scripts\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.940997 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-config-data\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941029 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-combined-ca-bundle\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941075 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-credential-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941108 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941145 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data-custom\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941170 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-fernet-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941199 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-internal-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941236 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-combined-ca-bundle\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941255 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98m6n\" (UniqueName: \"kubernetes.io/projected/f68b1099-7b2b-4d27-961c-ac412f438292-kube-api-access-98m6n\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.941282 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f68b1099-7b2b-4d27-961c-ac412f438292-logs\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.952411 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-96d8b7cb6-6rvkq"] Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.973512 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dk8s\" (UniqueName: \"kubernetes.io/projected/4c22d012-f519-4556-887a-2a0a11d7f3ec-kube-api-access-9dk8s\") pod \"placement-669f454766-khh4h\" (UID: \"4c22d012-f519-4556-887a-2a0a11d7f3ec\") " pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.977877 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-fernet-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.983898 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-config-data\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:40 crc kubenswrapper[4888]: I1124 00:45:40.993680 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-credential-keys\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.014228 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-scripts\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.014646 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-internal-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.015131 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-combined-ca-bundle\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.019092 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-public-tls-certs\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.019673 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrk6z\" (UniqueName: \"kubernetes.io/projected/b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10-kube-api-access-rrk6z\") pod \"keystone-5b7f898bdd-bbfm7\" (UID: \"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10\") " pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.043147 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-combined-ca-bundle\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.044289 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98m6n\" (UniqueName: \"kubernetes.io/projected/f68b1099-7b2b-4d27-961c-ac412f438292-kube-api-access-98m6n\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.044334 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f68b1099-7b2b-4d27-961c-ac412f438292-logs\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.044484 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.044514 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data-custom\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.045428 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f68b1099-7b2b-4d27-961c-ac412f438292-logs\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.053264 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.058679 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-combined-ca-bundle\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.069250 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f68b1099-7b2b-4d27-961c-ac412f438292-config-data-custom\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.101294 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98m6n\" (UniqueName: \"kubernetes.io/projected/f68b1099-7b2b-4d27-961c-ac412f438292-kube-api-access-98m6n\") pod \"barbican-keystone-listener-96d8b7cb6-6rvkq\" (UID: \"f68b1099-7b2b-4d27-961c-ac412f438292\") " pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.109677 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.118710 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-847b7f6b9f-ngtnz"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.133995 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.134910 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.136346 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.146616 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-847b7f6b9f-ngtnz"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.149155 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.159253 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247713 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247762 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data-custom\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247804 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247850 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247908 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247934 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-combined-ca-bundle\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247964 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76dg5\" (UniqueName: \"kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.247998 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.248039 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.248077 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bxp\" (UniqueName: \"kubernetes.io/projected/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-kube-api-access-m7bxp\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.248143 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-logs\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.270060 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.290951 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.292679 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.301661 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.313156 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.349996 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351127 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351278 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351399 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-combined-ca-bundle\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351537 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76dg5\" (UniqueName: \"kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351682 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351820 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.351923 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7bxp\" (UniqueName: \"kubernetes.io/projected/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-kube-api-access-m7bxp\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.352111 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-logs\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.352251 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.352370 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data-custom\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.354288 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.354841 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.356159 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-logs\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.356714 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.357163 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.357910 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.362618 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data-custom\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.366423 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-combined-ca-bundle\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.367119 4888 generic.go:334] "Generic (PLEG): container finished" podID="43659f95-de1a-41a7-b502-fd814dbac64a" containerID="e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac" exitCode=0 Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.367179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" event={"ID":"43659f95-de1a-41a7-b502-fd814dbac64a","Type":"ContainerDied","Data":"e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac"} Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.367205 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" event={"ID":"43659f95-de1a-41a7-b502-fd814dbac64a","Type":"ContainerDied","Data":"5a454c2f87eb69d7e3e980e152e27bcd42796e693afa07e8630a6b972f797a10"} Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.367225 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a454c2f87eb69d7e3e980e152e27bcd42796e693afa07e8630a6b972f797a10" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.370653 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.391733 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-config-data\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.393962 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76dg5\" (UniqueName: \"kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5\") pod \"dnsmasq-dns-85ff748b95-t52bs\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.413573 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7bxp\" (UniqueName: \"kubernetes.io/projected/b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf-kube-api-access-m7bxp\") pod \"barbican-worker-847b7f6b9f-ngtnz\" (UID: \"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf\") " pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.425528 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.454738 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrx6\" (UniqueName: \"kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.454789 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.454843 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.454893 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.454970 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.512885 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.546409 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556528 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkr44\" (UniqueName: \"kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556692 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556720 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556834 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.556851 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0\") pod \"43659f95-de1a-41a7-b502-fd814dbac64a\" (UID: \"43659f95-de1a-41a7-b502-fd814dbac64a\") " Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.557102 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.557121 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.557172 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.557237 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.557306 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrx6\" (UniqueName: \"kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.582132 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.582288 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44" (OuterVolumeSpecName: "kube-api-access-rkr44") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "kube-api-access-rkr44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.594597 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.605292 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrx6\" (UniqueName: \"kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.610073 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.611009 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data\") pod \"barbican-api-795b54f45d-r47pr\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.656886 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.659122 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkr44\" (UniqueName: \"kubernetes.io/projected/43659f95-de1a-41a7-b502-fd814dbac64a-kube-api-access-rkr44\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.704325 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.717881 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.733544 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config" (OuterVolumeSpecName: "config") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.743315 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.761412 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.761444 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.761453 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.761462 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.777971 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43659f95-de1a-41a7-b502-fd814dbac64a" (UID: "43659f95-de1a-41a7-b502-fd814dbac64a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.858084 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b7f898bdd-bbfm7"] Nov 24 00:45:41 crc kubenswrapper[4888]: I1124 00:45:41.863149 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43659f95-de1a-41a7-b502-fd814dbac64a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.208070 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-669f454766-khh4h"] Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.216366 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-96d8b7cb6-6rvkq"] Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.421447 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-847b7f6b9f-ngtnz"] Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.423042 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" event={"ID":"f68b1099-7b2b-4d27-961c-ac412f438292","Type":"ContainerStarted","Data":"99a72b6340372c8c4bc3b9937b8a4c770c5dfe1b35646f1544ec452c5562864d"} Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.426225 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-669f454766-khh4h" event={"ID":"4c22d012-f519-4556-887a-2a0a11d7f3ec","Type":"ContainerStarted","Data":"d004493d488c6df8bda1ec4222fd76d19766093bc1031d9171df623878652d40"} Nov 24 00:45:42 crc kubenswrapper[4888]: W1124 00:45:42.427093 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3c800fe_3dd6_47f6_b6b6_0dd825ed86bf.slice/crio-4f5104f982b185e362e2b0964cf90494f4ade7f537a7f2c5625dd7219227334f WatchSource:0}: Error finding container 4f5104f982b185e362e2b0964cf90494f4ade7f537a7f2c5625dd7219227334f: Status 404 returned error can't find the container with id 4f5104f982b185e362e2b0964cf90494f4ade7f537a7f2c5625dd7219227334f Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.429335 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-2v4q8" Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.431402 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b7f898bdd-bbfm7" event={"ID":"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10","Type":"ContainerStarted","Data":"463c99c77a03addaa2b069295a6336f2851a5c69806308a417747c8fb01da33b"} Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.431491 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b7f898bdd-bbfm7" event={"ID":"b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10","Type":"ContainerStarted","Data":"90192540f0798920d35f42abfd8532edf4ca2daefe2c6655f3279477b29e9165"} Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.431839 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.467544 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:45:42 crc kubenswrapper[4888]: W1124 00:45:42.478950 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod628a4a17_7824_4f16_ae50_eae75e8dd74e.slice/crio-b5a7ee4a3204715b283f90e20f7d1b70c84284a47a6f70e3b68ed7668a4cc160 WatchSource:0}: Error finding container b5a7ee4a3204715b283f90e20f7d1b70c84284a47a6f70e3b68ed7668a4cc160: Status 404 returned error can't find the container with id b5a7ee4a3204715b283f90e20f7d1b70c84284a47a6f70e3b68ed7668a4cc160 Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.507776 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b7f898bdd-bbfm7" podStartSLOduration=2.507761913 podStartE2EDuration="2.507761913s" podCreationTimestamp="2025-11-24 00:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:42.472641191 +0000 UTC m=+1245.055325245" watchObservedRunningTime="2025-11-24 00:45:42.507761913 +0000 UTC m=+1245.090445957" Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.529063 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.541304 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:42 crc kubenswrapper[4888]: I1124 00:45:42.560626 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-2v4q8"] Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.443852 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerStarted","Data":"122513ee9319822f26855ff6fd33bfcc578fe47adacf5019d562826bfc5505db"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.444297 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerStarted","Data":"b5a7ee4a3204715b283f90e20f7d1b70c84284a47a6f70e3b68ed7668a4cc160"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.446349 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-669f454766-khh4h" event={"ID":"4c22d012-f519-4556-887a-2a0a11d7f3ec","Type":"ContainerStarted","Data":"df389b472d6fb2140e7e57bc15f607cc77e435ecdf54c574a3a7ff0f4261afc8"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.446395 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-669f454766-khh4h" event={"ID":"4c22d012-f519-4556-887a-2a0a11d7f3ec","Type":"ContainerStarted","Data":"4636fcb654209adf3248b2914f5c6cde481bcbaf179837b6cf91b140d80a805c"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.446478 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.446722 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-669f454766-khh4h" Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.450094 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerStarted","Data":"eeb6a420b0489f318af2ff214fe609d3480e71cd65d9556ec8239b20a1a6b766"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.450134 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerStarted","Data":"5d020f64bf57bfe8ad02596ce5e051e569568536dec6107ee505dbac872e15df"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.454935 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" event={"ID":"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf","Type":"ContainerStarted","Data":"4f5104f982b185e362e2b0964cf90494f4ade7f537a7f2c5625dd7219227334f"} Nov 24 00:45:43 crc kubenswrapper[4888]: I1124 00:45:43.475259 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-669f454766-khh4h" podStartSLOduration=3.475244301 podStartE2EDuration="3.475244301s" podCreationTimestamp="2025-11-24 00:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:43.472536535 +0000 UTC m=+1246.055220589" watchObservedRunningTime="2025-11-24 00:45:43.475244301 +0000 UTC m=+1246.057928335" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.261931 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" path="/var/lib/kubelet/pods/43659f95-de1a-41a7-b502-fd814dbac64a/volumes" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.262804 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6dcbbbf466-gt2vf"] Nov 24 00:45:44 crc kubenswrapper[4888]: E1124 00:45:44.263257 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="init" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.263273 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="init" Nov 24 00:45:44 crc kubenswrapper[4888]: E1124 00:45:44.263297 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="dnsmasq-dns" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.263303 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="dnsmasq-dns" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.263481 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="43659f95-de1a-41a7-b502-fd814dbac64a" containerName="dnsmasq-dns" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.264727 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.267280 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.268016 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.280415 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dcbbbf466-gt2vf"] Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.318980 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pg8m\" (UniqueName: \"kubernetes.io/projected/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-kube-api-access-2pg8m\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319041 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-combined-ca-bundle\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319160 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data-custom\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319289 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-logs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319357 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-internal-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319608 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.319891 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-public-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.421469 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.421673 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-public-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.421761 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pg8m\" (UniqueName: \"kubernetes.io/projected/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-kube-api-access-2pg8m\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.421860 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-combined-ca-bundle\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.421975 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data-custom\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.422281 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-logs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.422357 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-internal-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.422866 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-logs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.425860 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-combined-ca-bundle\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.426331 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-public-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.426616 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.426789 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-config-data-custom\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.431013 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-internal-tls-certs\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.442483 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pg8m\" (UniqueName: \"kubernetes.io/projected/8e2c8c4b-730c-4c00-9dcd-03dd6b98b288-kube-api-access-2pg8m\") pod \"barbican-api-6dcbbbf466-gt2vf\" (UID: \"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288\") " pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.466237 4888 generic.go:334] "Generic (PLEG): container finished" podID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerID="eeb6a420b0489f318af2ff214fe609d3480e71cd65d9556ec8239b20a1a6b766" exitCode=0 Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.466311 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerDied","Data":"eeb6a420b0489f318af2ff214fe609d3480e71cd65d9556ec8239b20a1a6b766"} Nov 24 00:45:44 crc kubenswrapper[4888]: I1124 00:45:44.615854 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:45 crc kubenswrapper[4888]: I1124 00:45:45.179756 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dcbbbf466-gt2vf"] Nov 24 00:45:45 crc kubenswrapper[4888]: I1124 00:45:45.477146 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerStarted","Data":"5cdbaeed3e4e573d75d53f41e84ebc5b572cfb2467e33e63abc319f8bc4b1244"} Nov 24 00:45:45 crc kubenswrapper[4888]: I1124 00:45:45.478950 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2qlfb" event={"ID":"f5cc9373-09e2-4cae-82e0-badb0da1e0ad","Type":"ContainerStarted","Data":"a2f94b083b607c53a4955f05771ed86ec1873c0f906a2a4818209a87a66e83f1"} Nov 24 00:45:45 crc kubenswrapper[4888]: I1124 00:45:45.510425 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-795b54f45d-r47pr" podStartSLOduration=4.510409046 podStartE2EDuration="4.510409046s" podCreationTimestamp="2025-11-24 00:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:45.503260626 +0000 UTC m=+1248.085944700" watchObservedRunningTime="2025-11-24 00:45:45.510409046 +0000 UTC m=+1248.093093090" Nov 24 00:45:45 crc kubenswrapper[4888]: I1124 00:45:45.545539 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-2qlfb" podStartSLOduration=5.493265094 podStartE2EDuration="43.545519228s" podCreationTimestamp="2025-11-24 00:45:02 +0000 UTC" firstStartedPulling="2025-11-24 00:45:04.715603703 +0000 UTC m=+1207.298287747" lastFinishedPulling="2025-11-24 00:45:42.767857837 +0000 UTC m=+1245.350541881" observedRunningTime="2025-11-24 00:45:45.543876762 +0000 UTC m=+1248.126560806" watchObservedRunningTime="2025-11-24 00:45:45.545519228 +0000 UTC m=+1248.128203272" Nov 24 00:45:46 crc kubenswrapper[4888]: I1124 00:45:46.491502 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcbbbf466-gt2vf" event={"ID":"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288","Type":"ContainerStarted","Data":"44cbe856c0751241dd0910527b6020c78e8d458c606d52f09b38f596bb05b980"} Nov 24 00:45:46 crc kubenswrapper[4888]: I1124 00:45:46.491853 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:46 crc kubenswrapper[4888]: I1124 00:45:46.491872 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:49 crc kubenswrapper[4888]: I1124 00:45:49.523081 4888 generic.go:334] "Generic (PLEG): container finished" podID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" containerID="a2f94b083b607c53a4955f05771ed86ec1873c0f906a2a4818209a87a66e83f1" exitCode=0 Nov 24 00:45:49 crc kubenswrapper[4888]: I1124 00:45:49.523166 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2qlfb" event={"ID":"f5cc9373-09e2-4cae-82e0-badb0da1e0ad","Type":"ContainerDied","Data":"a2f94b083b607c53a4955f05771ed86ec1873c0f906a2a4818209a87a66e83f1"} Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.549376 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2qlfb" event={"ID":"f5cc9373-09e2-4cae-82e0-badb0da1e0ad","Type":"ContainerDied","Data":"15d73d15e5de8afae3a6ccff999c62eb270814b5e65db6df220be7c8e94a30c3"} Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.549703 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15d73d15e5de8afae3a6ccff999c62eb270814b5e65db6df220be7c8e94a30c3" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.579011 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.705213 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9452\" (UniqueName: \"kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452\") pod \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.705301 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data\") pod \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.705407 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle\") pod \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\" (UID: \"f5cc9373-09e2-4cae-82e0-badb0da1e0ad\") " Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.712664 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452" (OuterVolumeSpecName: "kube-api-access-k9452") pod "f5cc9373-09e2-4cae-82e0-badb0da1e0ad" (UID: "f5cc9373-09e2-4cae-82e0-badb0da1e0ad"). InnerVolumeSpecName "kube-api-access-k9452". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.786694 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5cc9373-09e2-4cae-82e0-badb0da1e0ad" (UID: "f5cc9373-09e2-4cae-82e0-badb0da1e0ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.801992 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data" (OuterVolumeSpecName: "config-data") pod "f5cc9373-09e2-4cae-82e0-badb0da1e0ad" (UID: "f5cc9373-09e2-4cae-82e0-badb0da1e0ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.807671 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.807699 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9452\" (UniqueName: \"kubernetes.io/projected/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-kube-api-access-k9452\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:51 crc kubenswrapper[4888]: I1124 00:45:51.807711 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc9373-09e2-4cae-82e0-badb0da1e0ad-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.597093 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerStarted","Data":"5991f032c722b6f92b9158aea88f2d0205f8231d91560b81c2445e208e85c180"} Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.598466 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.611214 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcbbbf466-gt2vf" event={"ID":"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288","Type":"ContainerStarted","Data":"3df4afd8b873c7fb2bc9b62f25bbcaf7602f246752089a7cd060e7d6ff1404eb"} Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.622771 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2qlfb" Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.623099 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" event={"ID":"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf","Type":"ContainerStarted","Data":"b052d7a40d6486ef3d898485078c13beaf3fb2836c9420b77d26759a05164951"} Nov 24 00:45:52 crc kubenswrapper[4888]: I1124 00:45:52.659595 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" podStartSLOduration=11.659574863 podStartE2EDuration="11.659574863s" podCreationTimestamp="2025-11-24 00:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:52.634739708 +0000 UTC m=+1255.217423752" watchObservedRunningTime="2025-11-24 00:45:52.659574863 +0000 UTC m=+1255.242258907" Nov 24 00:45:52 crc kubenswrapper[4888]: E1124 00:45:52.661271 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.145350 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.271719 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.591314 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.591401 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.637755 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" event={"ID":"b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf","Type":"ContainerStarted","Data":"b12293a258350f587a648b04b03a7c454582732d6e28d89aaa2d23a482e5ef0d"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.643644 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jzk9c" event={"ID":"1d5a8488-3848-4285-96bc-6313cc426ad5","Type":"ContainerStarted","Data":"7ddb707bef87a19cc9be039fbbf77a67ab65a2232d2d4f03a098472e123cd70e"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.659032 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerStarted","Data":"b91f45c28e7d15a1fc0ea29d4ce0a509486906df5a94a96f052eb8e7e6b4cc02"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.659092 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="ceilometer-notification-agent" containerID="cri-o://579a72c66538b5334ea3c608b33b8ac7e71f84d4d7adb9bdbecd081dcb6d21df" gracePeriod=30 Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.659198 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.659234 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="proxy-httpd" containerID="cri-o://b91f45c28e7d15a1fc0ea29d4ce0a509486906df5a94a96f052eb8e7e6b4cc02" gracePeriod=30 Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.659312 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="sg-core" containerID="cri-o://e698662ba24446ccef3f0cc1312c79317909ce77f9ef9c726d7141f83ed808e9" gracePeriod=30 Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.664225 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-847b7f6b9f-ngtnz" podStartSLOduration=3.979887694 podStartE2EDuration="13.664202559s" podCreationTimestamp="2025-11-24 00:45:40 +0000 UTC" firstStartedPulling="2025-11-24 00:45:42.438649021 +0000 UTC m=+1245.021333065" lastFinishedPulling="2025-11-24 00:45:52.122963886 +0000 UTC m=+1254.705647930" observedRunningTime="2025-11-24 00:45:53.655375232 +0000 UTC m=+1256.238059276" watchObservedRunningTime="2025-11-24 00:45:53.664202559 +0000 UTC m=+1256.246886603" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.675579 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" event={"ID":"f68b1099-7b2b-4d27-961c-ac412f438292","Type":"ContainerStarted","Data":"08e18b503368d2d87b194e6c068028785f7244dae7d5a63b2893a0d064b111cd"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.675629 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" event={"ID":"f68b1099-7b2b-4d27-961c-ac412f438292","Type":"ContainerStarted","Data":"c66c7362c57735366b0f30cfcb591b32094b4b88e9a9c9866ced28c17e774e05"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.684553 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jzk9c" podStartSLOduration=3.281046282 podStartE2EDuration="50.684536917s" podCreationTimestamp="2025-11-24 00:45:03 +0000 UTC" firstStartedPulling="2025-11-24 00:45:04.836262657 +0000 UTC m=+1207.418946701" lastFinishedPulling="2025-11-24 00:45:52.239753292 +0000 UTC m=+1254.822437336" observedRunningTime="2025-11-24 00:45:53.680406182 +0000 UTC m=+1256.263090226" watchObservedRunningTime="2025-11-24 00:45:53.684536917 +0000 UTC m=+1256.267220971" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.691370 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dcbbbf466-gt2vf" event={"ID":"8e2c8c4b-730c-4c00-9dcd-03dd6b98b288","Type":"ContainerStarted","Data":"45698ba3c13490c5789f412a14eae0a71c292e938d40dfd08be79a317667af27"} Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.692655 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.731494 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-96d8b7cb6-6rvkq" podStartSLOduration=3.8147465350000003 podStartE2EDuration="13.73147842s" podCreationTimestamp="2025-11-24 00:45:40 +0000 UTC" firstStartedPulling="2025-11-24 00:45:42.206170709 +0000 UTC m=+1244.788854753" lastFinishedPulling="2025-11-24 00:45:52.122902594 +0000 UTC m=+1254.705586638" observedRunningTime="2025-11-24 00:45:53.726635775 +0000 UTC m=+1256.309319819" watchObservedRunningTime="2025-11-24 00:45:53.73147842 +0000 UTC m=+1256.314162464" Nov 24 00:45:53 crc kubenswrapper[4888]: I1124 00:45:53.760881 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6dcbbbf466-gt2vf" podStartSLOduration=9.760864322 podStartE2EDuration="9.760864322s" podCreationTimestamp="2025-11-24 00:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:45:53.753352722 +0000 UTC m=+1256.336036766" watchObservedRunningTime="2025-11-24 00:45:53.760864322 +0000 UTC m=+1256.343548366" Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.616199 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705426 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerID="b91f45c28e7d15a1fc0ea29d4ce0a509486906df5a94a96f052eb8e7e6b4cc02" exitCode=0 Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705456 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerID="e698662ba24446ccef3f0cc1312c79317909ce77f9ef9c726d7141f83ed808e9" exitCode=2 Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705462 4888 generic.go:334] "Generic (PLEG): container finished" podID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerID="579a72c66538b5334ea3c608b33b8ac7e71f84d4d7adb9bdbecd081dcb6d21df" exitCode=0 Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705554 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerDied","Data":"b91f45c28e7d15a1fc0ea29d4ce0a509486906df5a94a96f052eb8e7e6b4cc02"} Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705606 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerDied","Data":"e698662ba24446ccef3f0cc1312c79317909ce77f9ef9c726d7141f83ed808e9"} Nov 24 00:45:54 crc kubenswrapper[4888]: I1124 00:45:54.705623 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerDied","Data":"579a72c66538b5334ea3c608b33b8ac7e71f84d4d7adb9bdbecd081dcb6d21df"} Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.171177 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.271905 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.272756 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.272837 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.272872 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.272921 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.273252 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flmff\" (UniqueName: \"kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.273643 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml\") pod \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\" (UID: \"ff46e22f-f4d5-414d-b152-a262ed0ab92c\") " Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.273804 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.273848 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.274370 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.274395 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff46e22f-f4d5-414d-b152-a262ed0ab92c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.285544 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff" (OuterVolumeSpecName: "kube-api-access-flmff") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "kube-api-access-flmff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.285926 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts" (OuterVolumeSpecName: "scripts") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.325319 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.331028 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.376740 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flmff\" (UniqueName: \"kubernetes.io/projected/ff46e22f-f4d5-414d-b152-a262ed0ab92c-kube-api-access-flmff\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.376780 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.376792 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.376804 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.378831 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data" (OuterVolumeSpecName: "config-data") pod "ff46e22f-f4d5-414d-b152-a262ed0ab92c" (UID: "ff46e22f-f4d5-414d-b152-a262ed0ab92c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.478880 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff46e22f-f4d5-414d-b152-a262ed0ab92c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.723613 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.724842 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff46e22f-f4d5-414d-b152-a262ed0ab92c","Type":"ContainerDied","Data":"46ca956f9dcd6d925d2b256de7efbb1c8fc287685c50cd69d5b6e7b205bc8112"} Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.725134 4888 scope.go:117] "RemoveContainer" containerID="b91f45c28e7d15a1fc0ea29d4ce0a509486906df5a94a96f052eb8e7e6b4cc02" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.750759 4888 scope.go:117] "RemoveContainer" containerID="e698662ba24446ccef3f0cc1312c79317909ce77f9ef9c726d7141f83ed808e9" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.777388 4888 scope.go:117] "RemoveContainer" containerID="579a72c66538b5334ea3c608b33b8ac7e71f84d4d7adb9bdbecd081dcb6d21df" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.815551 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.842190 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.863303 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:55 crc kubenswrapper[4888]: E1124 00:45:55.864074 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" containerName="heat-db-sync" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864091 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" containerName="heat-db-sync" Nov 24 00:45:55 crc kubenswrapper[4888]: E1124 00:45:55.864109 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="ceilometer-notification-agent" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864117 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="ceilometer-notification-agent" Nov 24 00:45:55 crc kubenswrapper[4888]: E1124 00:45:55.864134 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="sg-core" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864140 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="sg-core" Nov 24 00:45:55 crc kubenswrapper[4888]: E1124 00:45:55.864150 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="proxy-httpd" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864156 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="proxy-httpd" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864335 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="sg-core" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864352 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" containerName="heat-db-sync" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864369 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="proxy-httpd" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.864382 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" containerName="ceilometer-notification-agent" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.866537 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.868833 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.870944 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.876981 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.992981 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6gfm\" (UniqueName: \"kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993053 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993083 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993313 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993412 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993470 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:55 crc kubenswrapper[4888]: I1124 00:45:55.993746 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096172 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096222 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096266 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096307 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096332 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096471 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6gfm\" (UniqueName: \"kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.096839 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.097079 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.101994 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.102623 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.105242 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.106092 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.118120 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6gfm\" (UniqueName: \"kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm\") pod \"ceilometer-0\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.182748 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.261337 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff46e22f-f4d5-414d-b152-a262ed0ab92c" path="/var/lib/kubelet/pods/ff46e22f-f4d5-414d-b152-a262ed0ab92c/volumes" Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.651492 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:45:56 crc kubenswrapper[4888]: W1124 00:45:56.666888 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddafb43c7_844e_46f0_b94c_8c3aac9f0b6d.slice/crio-3307f98ef9203eea06fc6bf598cd42030c8ec51c60be7376c1871a135980e208 WatchSource:0}: Error finding container 3307f98ef9203eea06fc6bf598cd42030c8ec51c60be7376c1871a135980e208: Status 404 returned error can't find the container with id 3307f98ef9203eea06fc6bf598cd42030c8ec51c60be7376c1871a135980e208 Nov 24 00:45:56 crc kubenswrapper[4888]: I1124 00:45:56.733917 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerStarted","Data":"3307f98ef9203eea06fc6bf598cd42030c8ec51c60be7376c1871a135980e208"} Nov 24 00:45:57 crc kubenswrapper[4888]: I1124 00:45:57.748337 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerStarted","Data":"a3afaf059386d73e7c4f1ee441f9758e7849ba240a360d9ece19c1c7045a1a7b"} Nov 24 00:45:57 crc kubenswrapper[4888]: I1124 00:45:57.750349 4888 generic.go:334] "Generic (PLEG): container finished" podID="1d5a8488-3848-4285-96bc-6313cc426ad5" containerID="7ddb707bef87a19cc9be039fbbf77a67ab65a2232d2d4f03a098472e123cd70e" exitCode=0 Nov 24 00:45:57 crc kubenswrapper[4888]: I1124 00:45:57.750397 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jzk9c" event={"ID":"1d5a8488-3848-4285-96bc-6313cc426ad5","Type":"ContainerDied","Data":"7ddb707bef87a19cc9be039fbbf77a67ab65a2232d2d4f03a098472e123cd70e"} Nov 24 00:45:58 crc kubenswrapper[4888]: I1124 00:45:58.771913 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerStarted","Data":"9a537b028d2d3cec56d995f0ceae9e459ebdd004cc57721653b0a43d0a0f6fcc"} Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.221514 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.360874 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361064 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361126 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361146 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361177 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361203 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2kg9\" (UniqueName: \"kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9\") pod \"1d5a8488-3848-4285-96bc-6313cc426ad5\" (UID: \"1d5a8488-3848-4285-96bc-6313cc426ad5\") " Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361303 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.361796 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d5a8488-3848-4285-96bc-6313cc426ad5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.366950 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.367017 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9" (OuterVolumeSpecName: "kube-api-access-p2kg9") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "kube-api-access-p2kg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.368424 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts" (OuterVolumeSpecName: "scripts") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.398976 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.437136 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data" (OuterVolumeSpecName: "config-data") pod "1d5a8488-3848-4285-96bc-6313cc426ad5" (UID: "1d5a8488-3848-4285-96bc-6313cc426ad5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.463728 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.463764 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.463779 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.463791 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2kg9\" (UniqueName: \"kubernetes.io/projected/1d5a8488-3848-4285-96bc-6313cc426ad5-kube-api-access-p2kg9\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.463802 4888 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d5a8488-3848-4285-96bc-6313cc426ad5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.782029 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jzk9c" event={"ID":"1d5a8488-3848-4285-96bc-6313cc426ad5","Type":"ContainerDied","Data":"8bb9989d005364700bba60b490d3abb482a4cfa0d7dea27c87dbb267a33ad9f2"} Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.782600 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bb9989d005364700bba60b490d3abb482a4cfa0d7dea27c87dbb267a33ad9f2" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.782086 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jzk9c" Nov 24 00:45:59 crc kubenswrapper[4888]: I1124 00:45:59.784300 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerStarted","Data":"7f7c9697d077cf51363b1e8705d28aa75042e3a8c5485bb4835f0df4f74825f6"} Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.104430 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:00 crc kubenswrapper[4888]: E1124 00:46:00.104944 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" containerName="cinder-db-sync" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.104961 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" containerName="cinder-db-sync" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.105148 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" containerName="cinder-db-sync" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.106210 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.112943 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w7kcx" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.112987 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.113083 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.113131 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.131858 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.132112 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="dnsmasq-dns" containerID="cri-o://5991f032c722b6f92b9158aea88f2d0205f8231d91560b81c2445e208e85c180" gracePeriod=10 Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.146401 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.151591 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.177979 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.178030 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.178119 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.178153 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkcrx\" (UniqueName: \"kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.178202 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.178279 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.182781 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.184647 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.227606 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.284933 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285147 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285241 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285346 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285430 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkcrx\" (UniqueName: \"kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285532 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n2c5\" (UniqueName: \"kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285631 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285714 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285795 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.285911 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.286009 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.286111 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.286346 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.301691 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.302654 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.311725 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.320572 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkcrx\" (UniqueName: \"kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.321385 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.396459 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.396843 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n2c5\" (UniqueName: \"kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.396900 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.396932 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.396978 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.397042 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.398433 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.400330 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.400337 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.400456 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.402298 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.402395 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.410271 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.410416 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.425907 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.435147 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n2c5\" (UniqueName: \"kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5\") pod \"dnsmasq-dns-5c9776ccc5-clvbv\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.435373 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.488674 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498610 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498648 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498698 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498727 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498832 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmg8c\" (UniqueName: \"kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.498859 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.528458 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.600106 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.600659 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.600751 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.600900 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.600994 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.601420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.601658 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmg8c\" (UniqueName: \"kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.602009 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.604936 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.613057 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.614255 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.614903 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.617716 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.631318 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmg8c\" (UniqueName: \"kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c\") pod \"cinder-api-0\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.812201 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.850141 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerStarted","Data":"0aec7dd6106bfa49d067173e07a238ba214d214103b905dd9d140dc221ba581a"} Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.850861 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.856099 4888 generic.go:334] "Generic (PLEG): container finished" podID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerID="5991f032c722b6f92b9158aea88f2d0205f8231d91560b81c2445e208e85c180" exitCode=0 Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.856131 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerDied","Data":"5991f032c722b6f92b9158aea88f2d0205f8231d91560b81c2445e208e85c180"} Nov 24 00:46:00 crc kubenswrapper[4888]: I1124 00:46:00.953830 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.390730186 podStartE2EDuration="5.953801182s" podCreationTimestamp="2025-11-24 00:45:55 +0000 UTC" firstStartedPulling="2025-11-24 00:45:56.66963368 +0000 UTC m=+1259.252317724" lastFinishedPulling="2025-11-24 00:46:00.232704676 +0000 UTC m=+1262.815388720" observedRunningTime="2025-11-24 00:46:00.941162649 +0000 UTC m=+1263.523846693" watchObservedRunningTime="2025-11-24 00:46:00.953801182 +0000 UTC m=+1263.536485226" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.134561 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.184038 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:01 crc kubenswrapper[4888]: W1124 00:46:01.199300 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cb420c6_5c18_4cde_9e34_5dec80c3b72c.slice/crio-095cfe9239768b456acc4663f22e782d77307d38db7c661500b3335a1a8ba4a6 WatchSource:0}: Error finding container 095cfe9239768b456acc4663f22e782d77307d38db7c661500b3335a1a8ba4a6: Status 404 returned error can't find the container with id 095cfe9239768b456acc4663f22e782d77307d38db7c661500b3335a1a8ba4a6 Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220328 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220391 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220455 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76dg5\" (UniqueName: \"kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220501 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220533 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.220646 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb\") pod \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\" (UID: \"458edd3e-5009-41bf-b58f-0dc2d86f1a07\") " Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.253602 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.266092 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5" (OuterVolumeSpecName: "kube-api-access-76dg5") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "kube-api-access-76dg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.328378 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76dg5\" (UniqueName: \"kubernetes.io/projected/458edd3e-5009-41bf-b58f-0dc2d86f1a07-kube-api-access-76dg5\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.362849 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.370255 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.388706 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.390392 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config" (OuterVolumeSpecName: "config") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.425847 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "458edd3e-5009-41bf-b58f-0dc2d86f1a07" (UID: "458edd3e-5009-41bf-b58f-0dc2d86f1a07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.434394 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.434435 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.434448 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.434460 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.434472 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458edd3e-5009-41bf-b58f-0dc2d86f1a07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.606549 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.879977 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerStarted","Data":"401a846e8fb79f379d1f55d105fbdcda2d784935977137cfce133f8b6af03bd8"} Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.881769 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" event={"ID":"458edd3e-5009-41bf-b58f-0dc2d86f1a07","Type":"ContainerDied","Data":"5d020f64bf57bfe8ad02596ce5e051e569568536dec6107ee505dbac872e15df"} Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.881834 4888 scope.go:117] "RemoveContainer" containerID="5991f032c722b6f92b9158aea88f2d0205f8231d91560b81c2445e208e85c180" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.881957 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-t52bs" Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.883537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerStarted","Data":"543496c8d0064128d51b2b53950e00846c286acce99a364e239e64fd51c00a71"} Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.887037 4888 generic.go:334] "Generic (PLEG): container finished" podID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerID="145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574" exitCode=0 Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.887271 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" event={"ID":"6cb420c6-5c18-4cde-9e34-5dec80c3b72c","Type":"ContainerDied","Data":"145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574"} Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.887300 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" event={"ID":"6cb420c6-5c18-4cde-9e34-5dec80c3b72c","Type":"ContainerStarted","Data":"095cfe9239768b456acc4663f22e782d77307d38db7c661500b3335a1a8ba4a6"} Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.936877 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:46:01 crc kubenswrapper[4888]: I1124 00:46:01.948479 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-t52bs"] Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.005013 4888 scope.go:117] "RemoveContainer" containerID="eeb6a420b0489f318af2ff214fe609d3480e71cd65d9556ec8239b20a1a6b766" Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.269245 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" path="/var/lib/kubelet/pods/458edd3e-5009-41bf-b58f-0dc2d86f1a07/volumes" Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.334534 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.404562 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dcbbbf466-gt2vf" Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.479216 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.479469 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795b54f45d-r47pr" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api-log" containerID="cri-o://122513ee9319822f26855ff6fd33bfcc578fe47adacf5019d562826bfc5505db" gracePeriod=30 Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.479973 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795b54f45d-r47pr" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api" containerID="cri-o://5cdbaeed3e4e573d75d53f41e84ebc5b572cfb2467e33e63abc319f8bc4b1244" gracePeriod=30 Nov 24 00:46:02 crc kubenswrapper[4888]: I1124 00:46:02.872369 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-686d6d8cd9-m24dl" Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.015603 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" event={"ID":"6cb420c6-5c18-4cde-9e34-5dec80c3b72c","Type":"ContainerStarted","Data":"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a"} Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.015909 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.030298 4888 generic.go:334] "Generic (PLEG): container finished" podID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerID="122513ee9319822f26855ff6fd33bfcc578fe47adacf5019d562826bfc5505db" exitCode=143 Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.030384 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerDied","Data":"122513ee9319822f26855ff6fd33bfcc578fe47adacf5019d562826bfc5505db"} Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.033482 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.034055 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58bdd5575d-dsrnk" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-api" containerID="cri-o://989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7" gracePeriod=30 Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.034405 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58bdd5575d-dsrnk" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-httpd" containerID="cri-o://15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721" gracePeriod=30 Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.054056 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerStarted","Data":"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71"} Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.056072 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:03 crc kubenswrapper[4888]: I1124 00:46:03.060646 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" podStartSLOduration=3.060021135 podStartE2EDuration="3.060021135s" podCreationTimestamp="2025-11-24 00:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:03.039898863 +0000 UTC m=+1265.622582907" watchObservedRunningTime="2025-11-24 00:46:03.060021135 +0000 UTC m=+1265.642705179" Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.065854 4888 generic.go:334] "Generic (PLEG): container finished" podID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerID="15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721" exitCode=0 Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.065932 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerDied","Data":"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721"} Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.069527 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerStarted","Data":"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608"} Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.069663 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.069659 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api-log" containerID="cri-o://bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" gracePeriod=30 Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.069799 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api" containerID="cri-o://05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" gracePeriod=30 Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.072819 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerStarted","Data":"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b"} Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.072862 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerStarted","Data":"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534"} Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.099599 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.099582229 podStartE2EDuration="4.099582229s" podCreationTimestamp="2025-11-24 00:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:04.089720073 +0000 UTC m=+1266.672404117" watchObservedRunningTime="2025-11-24 00:46:04.099582229 +0000 UTC m=+1266.682266273" Nov 24 00:46:04 crc kubenswrapper[4888]: I1124 00:46:04.123647 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.174302232 podStartE2EDuration="4.1235996s" podCreationTimestamp="2025-11-24 00:46:00 +0000 UTC" firstStartedPulling="2025-11-24 00:46:01.187955551 +0000 UTC m=+1263.770639595" lastFinishedPulling="2025-11-24 00:46:02.137252919 +0000 UTC m=+1264.719936963" observedRunningTime="2025-11-24 00:46:04.115449892 +0000 UTC m=+1266.698133936" watchObservedRunningTime="2025-11-24 00:46:04.1235996 +0000 UTC m=+1266.706283644" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.078252 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086463 4888 generic.go:334] "Generic (PLEG): container finished" podID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerID="05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" exitCode=0 Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086495 4888 generic.go:334] "Generic (PLEG): container finished" podID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerID="bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" exitCode=143 Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086544 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerDied","Data":"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608"} Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086569 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086591 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerDied","Data":"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71"} Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086604 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80e1d83a-dabd-43dc-8860-1d2085b32fa1","Type":"ContainerDied","Data":"543496c8d0064128d51b2b53950e00846c286acce99a364e239e64fd51c00a71"} Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.086619 4888 scope.go:117] "RemoveContainer" containerID="05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.130932 4888 scope.go:117] "RemoveContainer" containerID="bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141476 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmg8c\" (UniqueName: \"kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141535 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141590 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141610 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141666 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141696 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.141726 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id\") pod \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\" (UID: \"80e1d83a-dabd-43dc-8860-1d2085b32fa1\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.142739 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs" (OuterVolumeSpecName: "logs") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.143376 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.148009 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.161953 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts" (OuterVolumeSpecName: "scripts") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.163257 4888 scope.go:117] "RemoveContainer" containerID="05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.168236 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c" (OuterVolumeSpecName: "kube-api-access-pmg8c") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "kube-api-access-pmg8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.168579 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608\": container with ID starting with 05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608 not found: ID does not exist" containerID="05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.168659 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608"} err="failed to get container status \"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608\": rpc error: code = NotFound desc = could not find container \"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608\": container with ID starting with 05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608 not found: ID does not exist" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.168692 4888 scope.go:117] "RemoveContainer" containerID="bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.169258 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71\": container with ID starting with bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71 not found: ID does not exist" containerID="bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.169319 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71"} err="failed to get container status \"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71\": rpc error: code = NotFound desc = could not find container \"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71\": container with ID starting with bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71 not found: ID does not exist" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.169339 4888 scope.go:117] "RemoveContainer" containerID="05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.171085 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608"} err="failed to get container status \"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608\": rpc error: code = NotFound desc = could not find container \"05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608\": container with ID starting with 05e2862982d7ae54eb36b837a13484b289be815c2cc398f01e9999f7f1a5a608 not found: ID does not exist" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.171115 4888 scope.go:117] "RemoveContainer" containerID="bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.171653 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71"} err="failed to get container status \"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71\": rpc error: code = NotFound desc = could not find container \"bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71\": container with ID starting with bac2c5a4a4a2986aab18e579bf481981d4c43a081880031d6b7f2db6a8631b71 not found: ID does not exist" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.191739 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.241927 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data" (OuterVolumeSpecName: "config-data") pod "80e1d83a-dabd-43dc-8860-1d2085b32fa1" (UID: "80e1d83a-dabd-43dc-8860-1d2085b32fa1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.244924 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmg8c\" (UniqueName: \"kubernetes.io/projected/80e1d83a-dabd-43dc-8860-1d2085b32fa1-kube-api-access-pmg8c\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245353 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245379 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245391 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245402 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80e1d83a-dabd-43dc-8860-1d2085b32fa1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245413 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80e1d83a-dabd-43dc-8860-1d2085b32fa1-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.245515 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80e1d83a-dabd-43dc-8860-1d2085b32fa1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.435874 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.473108 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.486012 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.494269 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.494747 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api-log" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.494772 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api-log" Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.494836 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="init" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.494846 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="init" Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.494872 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="dnsmasq-dns" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.494881 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="dnsmasq-dns" Nov 24 00:46:05 crc kubenswrapper[4888]: E1124 00:46:05.494902 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.494910 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.495151 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api-log" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.495171 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="458edd3e-5009-41bf-b58f-0dc2d86f1a07" containerName="dnsmasq-dns" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.495194 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" containerName="cinder-api" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.497832 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.504871 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.505678 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.505922 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.506012 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.557871 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558223 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2e8920-d187-4deb-85c8-b5b831af5995-logs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558244 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6lpq\" (UniqueName: \"kubernetes.io/projected/6a2e8920-d187-4deb-85c8-b5b831af5995-kube-api-access-v6lpq\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558263 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558281 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6a2e8920-d187-4deb-85c8-b5b831af5995-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558304 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558340 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558374 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data-custom\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.558421 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-scripts\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660020 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660448 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6a2e8920-d187-4deb-85c8-b5b831af5995-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660493 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660535 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660569 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data-custom\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660607 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-scripts\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660664 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660724 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2e8920-d187-4deb-85c8-b5b831af5995-logs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.660742 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6lpq\" (UniqueName: \"kubernetes.io/projected/6a2e8920-d187-4deb-85c8-b5b831af5995-kube-api-access-v6lpq\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.661971 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6a2e8920-d187-4deb-85c8-b5b831af5995-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.664302 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a2e8920-d187-4deb-85c8-b5b831af5995-logs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.664700 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.665957 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-scripts\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.666462 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.668003 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-config-data-custom\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.668324 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.669126 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a2e8920-d187-4deb-85c8-b5b831af5995-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.677151 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6lpq\" (UniqueName: \"kubernetes.io/projected/6a2e8920-d187-4deb-85c8-b5b831af5995-kube-api-access-v6lpq\") pod \"cinder-api-0\" (UID: \"6a2e8920-d187-4deb-85c8-b5b831af5995\") " pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.835845 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.877973 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.965524 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs\") pod \"96baab99-3e6b-4035-85ed-acaed5a4c96f\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.965969 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle\") pod \"96baab99-3e6b-4035-85ed-acaed5a4c96f\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.966184 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-466kt\" (UniqueName: \"kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt\") pod \"96baab99-3e6b-4035-85ed-acaed5a4c96f\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.966247 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config\") pod \"96baab99-3e6b-4035-85ed-acaed5a4c96f\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.966295 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config\") pod \"96baab99-3e6b-4035-85ed-acaed5a4c96f\" (UID: \"96baab99-3e6b-4035-85ed-acaed5a4c96f\") " Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.976072 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "96baab99-3e6b-4035-85ed-acaed5a4c96f" (UID: "96baab99-3e6b-4035-85ed-acaed5a4c96f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:05 crc kubenswrapper[4888]: I1124 00:46:05.988552 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt" (OuterVolumeSpecName: "kube-api-access-466kt") pod "96baab99-3e6b-4035-85ed-acaed5a4c96f" (UID: "96baab99-3e6b-4035-85ed-acaed5a4c96f"). InnerVolumeSpecName "kube-api-access-466kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.057190 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config" (OuterVolumeSpecName: "config") pod "96baab99-3e6b-4035-85ed-acaed5a4c96f" (UID: "96baab99-3e6b-4035-85ed-acaed5a4c96f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.059530 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96baab99-3e6b-4035-85ed-acaed5a4c96f" (UID: "96baab99-3e6b-4035-85ed-acaed5a4c96f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.068845 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.068878 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.068892 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-466kt\" (UniqueName: \"kubernetes.io/projected/96baab99-3e6b-4035-85ed-acaed5a4c96f-kube-api-access-466kt\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.068903 4888 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.089596 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "96baab99-3e6b-4035-85ed-acaed5a4c96f" (UID: "96baab99-3e6b-4035-85ed-acaed5a4c96f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.098231 4888 generic.go:334] "Generic (PLEG): container finished" podID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerID="5cdbaeed3e4e573d75d53f41e84ebc5b572cfb2467e33e63abc319f8bc4b1244" exitCode=0 Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.098310 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerDied","Data":"5cdbaeed3e4e573d75d53f41e84ebc5b572cfb2467e33e63abc319f8bc4b1244"} Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.100411 4888 generic.go:334] "Generic (PLEG): container finished" podID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerID="989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7" exitCode=0 Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.100481 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerDied","Data":"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7"} Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.100512 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58bdd5575d-dsrnk" event={"ID":"96baab99-3e6b-4035-85ed-acaed5a4c96f","Type":"ContainerDied","Data":"17d6c2e265b22641759208996e1b06b56a1158709814ad133cd9dddf69ae60d8"} Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.100532 4888 scope.go:117] "RemoveContainer" containerID="15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.100646 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58bdd5575d-dsrnk" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.135385 4888 scope.go:117] "RemoveContainer" containerID="989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.143391 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.150553 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-58bdd5575d-dsrnk"] Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.168019 4888 scope.go:117] "RemoveContainer" containerID="15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721" Nov 24 00:46:06 crc kubenswrapper[4888]: E1124 00:46:06.168496 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721\": container with ID starting with 15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721 not found: ID does not exist" containerID="15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.168535 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721"} err="failed to get container status \"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721\": rpc error: code = NotFound desc = could not find container \"15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721\": container with ID starting with 15a177a05222d698fa7a70184e661696069b97c99304507e7b7c1b99dbb26721 not found: ID does not exist" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.168597 4888 scope.go:117] "RemoveContainer" containerID="989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7" Nov 24 00:46:06 crc kubenswrapper[4888]: E1124 00:46:06.169222 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7\": container with ID starting with 989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7 not found: ID does not exist" containerID="989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.169303 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7"} err="failed to get container status \"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7\": rpc error: code = NotFound desc = could not find container \"989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7\": container with ID starting with 989351a9275b572502492f5bff7d3f24f766f13b88ff86af15b37fa8b3ef0ff7 not found: ID does not exist" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.170701 4888 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96baab99-3e6b-4035-85ed-acaed5a4c96f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.291626 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e1d83a-dabd-43dc-8860-1d2085b32fa1" path="/var/lib/kubelet/pods/80e1d83a-dabd-43dc-8860-1d2085b32fa1/volumes" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.292543 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" path="/var/lib/kubelet/pods/96baab99-3e6b-4035-85ed-acaed5a4c96f/volumes" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.323386 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.466160 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.587622 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxrx6\" (UniqueName: \"kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6\") pod \"628a4a17-7824-4f16-ae50-eae75e8dd74e\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.587826 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs\") pod \"628a4a17-7824-4f16-ae50-eae75e8dd74e\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.587943 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data\") pod \"628a4a17-7824-4f16-ae50-eae75e8dd74e\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.588022 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom\") pod \"628a4a17-7824-4f16-ae50-eae75e8dd74e\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.588090 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle\") pod \"628a4a17-7824-4f16-ae50-eae75e8dd74e\" (UID: \"628a4a17-7824-4f16-ae50-eae75e8dd74e\") " Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.588265 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs" (OuterVolumeSpecName: "logs") pod "628a4a17-7824-4f16-ae50-eae75e8dd74e" (UID: "628a4a17-7824-4f16-ae50-eae75e8dd74e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.588633 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/628a4a17-7824-4f16-ae50-eae75e8dd74e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.592133 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6" (OuterVolumeSpecName: "kube-api-access-sxrx6") pod "628a4a17-7824-4f16-ae50-eae75e8dd74e" (UID: "628a4a17-7824-4f16-ae50-eae75e8dd74e"). InnerVolumeSpecName "kube-api-access-sxrx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.596058 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "628a4a17-7824-4f16-ae50-eae75e8dd74e" (UID: "628a4a17-7824-4f16-ae50-eae75e8dd74e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.618353 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "628a4a17-7824-4f16-ae50-eae75e8dd74e" (UID: "628a4a17-7824-4f16-ae50-eae75e8dd74e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.656477 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data" (OuterVolumeSpecName: "config-data") pod "628a4a17-7824-4f16-ae50-eae75e8dd74e" (UID: "628a4a17-7824-4f16-ae50-eae75e8dd74e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.691120 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.691154 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxrx6\" (UniqueName: \"kubernetes.io/projected/628a4a17-7824-4f16-ae50-eae75e8dd74e-kube-api-access-sxrx6\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.691196 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:06 crc kubenswrapper[4888]: I1124 00:46:06.691211 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/628a4a17-7824-4f16-ae50-eae75e8dd74e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.120752 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795b54f45d-r47pr" event={"ID":"628a4a17-7824-4f16-ae50-eae75e8dd74e","Type":"ContainerDied","Data":"b5a7ee4a3204715b283f90e20f7d1b70c84284a47a6f70e3b68ed7668a4cc160"} Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.120770 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795b54f45d-r47pr" Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.121092 4888 scope.go:117] "RemoveContainer" containerID="5cdbaeed3e4e573d75d53f41e84ebc5b572cfb2467e33e63abc319f8bc4b1244" Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.132421 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6a2e8920-d187-4deb-85c8-b5b831af5995","Type":"ContainerStarted","Data":"002954c4a41b3d0d9b264dd14a8e1a2ef06863dba907fee658bf3021c26a541b"} Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.132473 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6a2e8920-d187-4deb-85c8-b5b831af5995","Type":"ContainerStarted","Data":"2f752cc6092c91dd339e40b76a2e9d44ace99eccf39c3bece03e4dcee1134995"} Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.158545 4888 scope.go:117] "RemoveContainer" containerID="122513ee9319822f26855ff6fd33bfcc578fe47adacf5019d562826bfc5505db" Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.161066 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:46:07 crc kubenswrapper[4888]: I1124 00:46:07.171956 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-795b54f45d-r47pr"] Nov 24 00:46:08 crc kubenswrapper[4888]: I1124 00:46:08.150869 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6a2e8920-d187-4deb-85c8-b5b831af5995","Type":"ContainerStarted","Data":"13d12ed1c3fefbade9763b671d11a17054d020eedbd4391a0cacaaf7cd1045ef"} Nov 24 00:46:08 crc kubenswrapper[4888]: I1124 00:46:08.151454 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 00:46:08 crc kubenswrapper[4888]: I1124 00:46:08.176762 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.176738692 podStartE2EDuration="3.176738692s" podCreationTimestamp="2025-11-24 00:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:08.171610189 +0000 UTC m=+1270.754294233" watchObservedRunningTime="2025-11-24 00:46:08.176738692 +0000 UTC m=+1270.759422756" Nov 24 00:46:08 crc kubenswrapper[4888]: I1124 00:46:08.267477 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" path="/var/lib/kubelet/pods/628a4a17-7824-4f16-ae50-eae75e8dd74e/volumes" Nov 24 00:46:10 crc kubenswrapper[4888]: I1124 00:46:10.492100 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:10 crc kubenswrapper[4888]: I1124 00:46:10.628535 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:46:10 crc kubenswrapper[4888]: I1124 00:46:10.628789 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="dnsmasq-dns" containerID="cri-o://637fe1043e79450f555a42fea83447bf5def0711407e00f87b4463a927affed5" gracePeriod=10 Nov 24 00:46:10 crc kubenswrapper[4888]: I1124 00:46:10.773374 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 00:46:10 crc kubenswrapper[4888]: I1124 00:46:10.830731 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203045 4888 generic.go:334] "Generic (PLEG): container finished" podID="4f2220ab-7482-436b-8386-689c044405b3" containerID="637fe1043e79450f555a42fea83447bf5def0711407e00f87b4463a927affed5" exitCode=0 Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203249 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="cinder-scheduler" containerID="cri-o://269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534" gracePeriod=30 Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203356 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" event={"ID":"4f2220ab-7482-436b-8386-689c044405b3","Type":"ContainerDied","Data":"637fe1043e79450f555a42fea83447bf5def0711407e00f87b4463a927affed5"} Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203406 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" event={"ID":"4f2220ab-7482-436b-8386-689c044405b3","Type":"ContainerDied","Data":"1d5d54aa3a97f9459f78504704ab297b9f40332817621c44629a170877228e2d"} Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203418 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d5d54aa3a97f9459f78504704ab297b9f40332817621c44629a170877228e2d" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.203363 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="probe" containerID="cri-o://960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b" gracePeriod=30 Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.256424 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.391405 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwsgv\" (UniqueName: \"kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.392858 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.393125 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.393293 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.393413 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.393628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config\") pod \"4f2220ab-7482-436b-8386-689c044405b3\" (UID: \"4f2220ab-7482-436b-8386-689c044405b3\") " Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.398003 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv" (OuterVolumeSpecName: "kube-api-access-zwsgv") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "kube-api-access-zwsgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.474272 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config" (OuterVolumeSpecName: "config") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.492307 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.498212 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwsgv\" (UniqueName: \"kubernetes.io/projected/4f2220ab-7482-436b-8386-689c044405b3-kube-api-access-zwsgv\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.499106 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.499166 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.500377 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.503372 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.540351 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f2220ab-7482-436b-8386-689c044405b3" (UID: "4f2220ab-7482-436b-8386-689c044405b3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.600654 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.600686 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: I1124 00:46:11.600696 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f2220ab-7482-436b-8386-689c044405b3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:11 crc kubenswrapper[4888]: E1124 00:46:11.690961 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87029c39_e167_4ff0_ac35_e04144b94835.slice/crio-960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.215463 4888 generic.go:334] "Generic (PLEG): container finished" podID="87029c39-e167-4ff0-ac35-e04144b94835" containerID="960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b" exitCode=0 Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.215598 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerDied","Data":"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b"} Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.215885 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-bpqx2" Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.287925 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.287974 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-bpqx2"] Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.391048 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-669f454766-khh4h" Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.393099 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-669f454766-khh4h" Nov 24 00:46:12 crc kubenswrapper[4888]: I1124 00:46:12.909051 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b7f898bdd-bbfm7" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.257864 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f2220ab-7482-436b-8386-689c044405b3" path="/var/lib/kubelet/pods/4f2220ab-7482-436b-8386-689c044405b3/volumes" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.813265 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.860944 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.861040 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.861077 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.861159 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.861252 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.861274 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkcrx\" (UniqueName: \"kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx\") pod \"87029c39-e167-4ff0-ac35-e04144b94835\" (UID: \"87029c39-e167-4ff0-ac35-e04144b94835\") " Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.864061 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.868108 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx" (OuterVolumeSpecName: "kube-api-access-xkcrx") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "kube-api-access-xkcrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.868763 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts" (OuterVolumeSpecName: "scripts") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.877850 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.927045 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.964204 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.964618 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.964685 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.964741 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87029c39-e167-4ff0-ac35-e04144b94835-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.964893 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkcrx\" (UniqueName: \"kubernetes.io/projected/87029c39-e167-4ff0-ac35-e04144b94835-kube-api-access-xkcrx\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:14 crc kubenswrapper[4888]: I1124 00:46:14.993324 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data" (OuterVolumeSpecName: "config-data") pod "87029c39-e167-4ff0-ac35-e04144b94835" (UID: "87029c39-e167-4ff0-ac35-e04144b94835"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.067141 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87029c39-e167-4ff0-ac35-e04144b94835-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.244031 4888 generic.go:334] "Generic (PLEG): container finished" podID="87029c39-e167-4ff0-ac35-e04144b94835" containerID="269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534" exitCode=0 Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.244106 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerDied","Data":"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534"} Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.244145 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87029c39-e167-4ff0-ac35-e04144b94835","Type":"ContainerDied","Data":"401a846e8fb79f379d1f55d105fbdcda2d784935977137cfce133f8b6af03bd8"} Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.244164 4888 scope.go:117] "RemoveContainer" containerID="960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.244125 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.277264 4888 scope.go:117] "RemoveContainer" containerID="269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.278340 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.288338 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.300809 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.301484 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.301588 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.301669 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="dnsmasq-dns" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.301740 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="dnsmasq-dns" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.301816 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api-log" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.301911 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api-log" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.302009 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="init" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.302091 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="init" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.302175 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="probe" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.302276 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="probe" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.302367 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-api" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.302434 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-api" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.302517 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="cinder-scheduler" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.302591 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="cinder-scheduler" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.302663 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-httpd" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.302728 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-httpd" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303028 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-httpd" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303114 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api-log" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303219 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="96baab99-3e6b-4035-85ed-acaed5a4c96f" containerName="neutron-api" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303339 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="628a4a17-7824-4f16-ae50-eae75e8dd74e" containerName="barbican-api" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303429 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="probe" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303512 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="87029c39-e167-4ff0-ac35-e04144b94835" containerName="cinder-scheduler" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.303585 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f2220ab-7482-436b-8386-689c044405b3" containerName="dnsmasq-dns" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.304928 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.308880 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.318225 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.318426 4888 scope.go:117] "RemoveContainer" containerID="960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.319371 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b\": container with ID starting with 960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b not found: ID does not exist" containerID="960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.319513 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b"} err="failed to get container status \"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b\": rpc error: code = NotFound desc = could not find container \"960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b\": container with ID starting with 960a06e0f134399b1270cbca420789426825210f608df4e804f369f9deb4845b not found: ID does not exist" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.319630 4888 scope.go:117] "RemoveContainer" containerID="269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534" Nov 24 00:46:15 crc kubenswrapper[4888]: E1124 00:46:15.320043 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534\": container with ID starting with 269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534 not found: ID does not exist" containerID="269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.320084 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534"} err="failed to get container status \"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534\": rpc error: code = NotFound desc = could not find container \"269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534\": container with ID starting with 269401313a4d666bad035f578d6bc696d913dc4681fc218e048553a060f69534 not found: ID does not exist" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.382558 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.382628 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.382858 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-scripts\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.382903 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d98628-e960-4e56-b620-7ba6050faadf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.382984 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.383057 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjtw\" (UniqueName: \"kubernetes.io/projected/32d98628-e960-4e56-b620-7ba6050faadf-kube-api-access-4jjtw\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjtw\" (UniqueName: \"kubernetes.io/projected/32d98628-e960-4e56-b620-7ba6050faadf-kube-api-access-4jjtw\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485633 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485702 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-scripts\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485725 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d98628-e960-4e56-b620-7ba6050faadf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.485751 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.486407 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d98628-e960-4e56-b620-7ba6050faadf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.490496 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.490608 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.490777 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-scripts\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.492674 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d98628-e960-4e56-b620-7ba6050faadf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.511637 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjtw\" (UniqueName: \"kubernetes.io/projected/32d98628-e960-4e56-b620-7ba6050faadf-kube-api-access-4jjtw\") pod \"cinder-scheduler-0\" (UID: \"32d98628-e960-4e56-b620-7ba6050faadf\") " pod="openstack/cinder-scheduler-0" Nov 24 00:46:15 crc kubenswrapper[4888]: I1124 00:46:15.627965 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.093452 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 00:46:16 crc kubenswrapper[4888]: W1124 00:46:16.099149 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32d98628_e960_4e56_b620_7ba6050faadf.slice/crio-a1647a952d490511309ac119102e545cebfedf28bf200c551443bd3550a17222 WatchSource:0}: Error finding container a1647a952d490511309ac119102e545cebfedf28bf200c551443bd3550a17222: Status 404 returned error can't find the container with id a1647a952d490511309ac119102e545cebfedf28bf200c551443bd3550a17222 Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.267087 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87029c39-e167-4ff0-ac35-e04144b94835" path="/var/lib/kubelet/pods/87029c39-e167-4ff0-ac35-e04144b94835/volumes" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.284724 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"32d98628-e960-4e56-b620-7ba6050faadf","Type":"ContainerStarted","Data":"a1647a952d490511309ac119102e545cebfedf28bf200c551443bd3550a17222"} Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.366436 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.368034 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.382045 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.382238 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tt2jh" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.382486 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.414484 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.483938 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.485723 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.509560 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jndxh\" (UniqueName: \"kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.509633 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.509714 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.509758 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.528214 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.553407 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.555120 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.558297 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.562231 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.574273 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.580171 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.588086 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.594949 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614703 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614783 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614846 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nrk7\" (UniqueName: \"kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614889 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614918 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.614954 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615004 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615047 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615072 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jndxh\" (UniqueName: \"kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615146 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615174 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615211 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615281 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.615320 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2lg\" (UniqueName: \"kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.624312 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.628981 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.651299 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.663295 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jndxh\" (UniqueName: \"kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh\") pod \"heat-engine-79868646c7-2h6p7\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716747 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdx76\" (UniqueName: \"kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716808 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716857 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nrk7\" (UniqueName: \"kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716906 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716940 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716962 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.716990 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717017 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717044 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717092 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717115 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717140 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717182 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717206 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2lg\" (UniqueName: \"kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.717676 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.718280 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.718478 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.719263 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.719676 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.724675 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.730698 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.730847 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.735239 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nrk7\" (UniqueName: \"kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7\") pod \"dnsmasq-dns-7756b9d78c-brzj4\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.735296 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.742834 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2lg\" (UniqueName: \"kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg\") pod \"heat-cfnapi-7c4dff5d5b-9qwfl\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.819562 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.819713 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdx76\" (UniqueName: \"kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.819770 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.819878 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.823401 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.824497 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.825863 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.838401 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.840704 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdx76\" (UniqueName: \"kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76\") pod \"heat-api-559fcf88cb-n5qbg\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:16 crc kubenswrapper[4888]: I1124 00:46:16.898941 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.133672 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.198732 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6dd96dd9bc-c9x6n"] Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.200430 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.203590 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.203792 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.204381 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.207891 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6dd96dd9bc-c9x6n"] Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.323980 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"32d98628-e960-4e56-b620-7ba6050faadf","Type":"ContainerStarted","Data":"3d8f1f2daf02ae0888fdf8cedcc909cc5a628ea19153a2b6798b9eaf9960a789"} Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.341832 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-public-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.341922 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24x4h\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-kube-api-access-24x4h\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342009 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-etc-swift\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342049 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-log-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342070 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-run-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342123 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-combined-ca-bundle\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342193 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-config-data\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.342273 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-internal-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.364431 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-internal-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456472 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-public-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456503 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24x4h\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-kube-api-access-24x4h\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456540 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-etc-swift\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456571 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-log-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456595 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-run-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456629 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-combined-ca-bundle\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.456679 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-config-data\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.457453 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-run-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.457716 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07c3d65f-a35c-481d-a773-a9dc4dd914b5-log-httpd\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.467929 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-etc-swift\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.481720 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-public-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.485772 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-combined-ca-bundle\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.488273 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-internal-tls-certs\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.503522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24x4h\" (UniqueName: \"kubernetes.io/projected/07c3d65f-a35c-481d-a773-a9dc4dd914b5-kube-api-access-24x4h\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.527583 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c3d65f-a35c-481d-a773-a9dc4dd914b5-config-data\") pod \"swift-proxy-6dd96dd9bc-c9x6n\" (UID: \"07c3d65f-a35c-481d-a773-a9dc4dd914b5\") " pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.530643 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.556388 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.563184 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.578533 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qfwnq" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.578765 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.578954 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.582221 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.665782 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkm2j\" (UniqueName: \"kubernetes.io/projected/a036e658-146d-4eb0-a197-939e6bbc31f8-kube-api-access-wkm2j\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.666209 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.666248 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.666287 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.768990 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.769061 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.769092 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkm2j\" (UniqueName: \"kubernetes.io/projected/a036e658-146d-4eb0-a197-939e6bbc31f8-kube-api-access-wkm2j\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.769205 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.770064 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.782615 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.782992 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a036e658-146d-4eb0-a197-939e6bbc31f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.797882 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkm2j\" (UniqueName: \"kubernetes.io/projected/a036e658-146d-4eb0-a197-939e6bbc31f8-kube-api-access-wkm2j\") pod \"openstackclient\" (UID: \"a036e658-146d-4eb0-a197-939e6bbc31f8\") " pod="openstack/openstackclient" Nov 24 00:46:17 crc kubenswrapper[4888]: I1124 00:46:17.979689 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.057712 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.092157 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.274663 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.370141 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-559fcf88cb-n5qbg" event={"ID":"eb2f6347-c0a7-4af3-b9f2-7df93d807a88","Type":"ContainerStarted","Data":"a5dfec0fa404e06253ab457943f0f32363b2b4fb222c615b305691e5a126c5c5"} Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.377661 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" event={"ID":"c9b000bb-1961-4ece-b8df-13dd16d4e2fa","Type":"ContainerStarted","Data":"1c9643c47170e49bf0df0067e3ac4755a5b1b69b0f4e554752da5b631b1993b4"} Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.428497 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6dd96dd9bc-c9x6n"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.442575 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-79868646c7-2h6p7" event={"ID":"241b6970-6ac3-49b0-8b51-8f727991f324","Type":"ContainerStarted","Data":"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875"} Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.442609 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-79868646c7-2h6p7" event={"ID":"241b6970-6ac3-49b0-8b51-8f727991f324","Type":"ContainerStarted","Data":"4dd985148283d4a5e153768edecedcb2ac5c81b750a7c97d49b5092a74d9b03f"} Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.442730 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.445585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" event={"ID":"a481e8b8-30cb-4449-99fe-24929099c3b5","Type":"ContainerStarted","Data":"3e30ebe604ce6d231b0fef04941f9d694558f32abf5b01ad9f9f00f56b621327"} Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.458452 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-79868646c7-2h6p7" podStartSLOduration=2.458430984 podStartE2EDuration="2.458430984s" podCreationTimestamp="2025-11-24 00:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:18.458248199 +0000 UTC m=+1281.040932263" watchObservedRunningTime="2025-11-24 00:46:18.458430984 +0000 UTC m=+1281.041115028" Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.642548 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.688523 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.688846 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-central-agent" containerID="cri-o://a3afaf059386d73e7c4f1ee441f9758e7849ba240a360d9ece19c1c7045a1a7b" gracePeriod=30 Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.689585 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="proxy-httpd" containerID="cri-o://0aec7dd6106bfa49d067173e07a238ba214d214103b905dd9d140dc221ba581a" gracePeriod=30 Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.689638 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="sg-core" containerID="cri-o://7f7c9697d077cf51363b1e8705d28aa75042e3a8c5485bb4835f0df4f74825f6" gracePeriod=30 Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.689678 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-notification-agent" containerID="cri-o://9a537b028d2d3cec56d995f0ceae9e459ebdd004cc57721653b0a43d0a0f6fcc" gracePeriod=30 Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.690658 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 00:46:18 crc kubenswrapper[4888]: I1124 00:46:18.702033 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.476055 4888 generic.go:334] "Generic (PLEG): container finished" podID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerID="e30aa68d1d49594dbedebf3412425956f6a8c49ad68565b4b75d66881f24a1e1" exitCode=0 Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.476255 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" event={"ID":"c9b000bb-1961-4ece-b8df-13dd16d4e2fa","Type":"ContainerDied","Data":"e30aa68d1d49594dbedebf3412425956f6a8c49ad68565b4b75d66881f24a1e1"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.479334 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"32d98628-e960-4e56-b620-7ba6050faadf","Type":"ContainerStarted","Data":"29d16ea4c64daa9c15756e6ec1c8fa5d1014e5a97a517a24174cf9e7b76bd1d9"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.485016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a036e658-146d-4eb0-a197-939e6bbc31f8","Type":"ContainerStarted","Data":"14441ab13bc48d77a5751a86094a9560c49e39eb261f4f67eb12d4477a977dbf"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.489850 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" event={"ID":"07c3d65f-a35c-481d-a773-a9dc4dd914b5","Type":"ContainerStarted","Data":"011933068e9478adf5c39e68cda089eadacc00405b6cf3ea54c8556f3c099cbf"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.489897 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" event={"ID":"07c3d65f-a35c-481d-a773-a9dc4dd914b5","Type":"ContainerStarted","Data":"5a05e1b6468985ad87c86961f40bdafdbe54c09c58ff671be7a15268b06b4442"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.489907 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" event={"ID":"07c3d65f-a35c-481d-a773-a9dc4dd914b5","Type":"ContainerStarted","Data":"c4eb543bf69b7b40eb27d3c74e0029d3aaa312285f53d5b9839c650eca53b4d7"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.492054 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.492087 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.502718 4888 generic.go:334] "Generic (PLEG): container finished" podID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerID="0aec7dd6106bfa49d067173e07a238ba214d214103b905dd9d140dc221ba581a" exitCode=0 Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.502743 4888 generic.go:334] "Generic (PLEG): container finished" podID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerID="7f7c9697d077cf51363b1e8705d28aa75042e3a8c5485bb4835f0df4f74825f6" exitCode=2 Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.502751 4888 generic.go:334] "Generic (PLEG): container finished" podID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerID="a3afaf059386d73e7c4f1ee441f9758e7849ba240a360d9ece19c1c7045a1a7b" exitCode=0 Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.503558 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerDied","Data":"0aec7dd6106bfa49d067173e07a238ba214d214103b905dd9d140dc221ba581a"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.503589 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerDied","Data":"7f7c9697d077cf51363b1e8705d28aa75042e3a8c5485bb4835f0df4f74825f6"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.503600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerDied","Data":"a3afaf059386d73e7c4f1ee441f9758e7849ba240a360d9ece19c1c7045a1a7b"} Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.563286 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.5632505420000005 podStartE2EDuration="4.563250542s" podCreationTimestamp="2025-11-24 00:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:19.519889499 +0000 UTC m=+1282.102573543" watchObservedRunningTime="2025-11-24 00:46:19.563250542 +0000 UTC m=+1282.145934586" Nov 24 00:46:19 crc kubenswrapper[4888]: I1124 00:46:19.591904 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" podStartSLOduration=2.591876192 podStartE2EDuration="2.591876192s" podCreationTimestamp="2025-11-24 00:46:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:19.545215927 +0000 UTC m=+1282.127899981" watchObservedRunningTime="2025-11-24 00:46:19.591876192 +0000 UTC m=+1282.174560276" Nov 24 00:46:20 crc kubenswrapper[4888]: I1124 00:46:20.539418 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" event={"ID":"c9b000bb-1961-4ece-b8df-13dd16d4e2fa","Type":"ContainerStarted","Data":"e9c8478655a21127b52ab79ea30d43fc2ca3a9c4f02745a8f438b709b73f0f3a"} Nov 24 00:46:20 crc kubenswrapper[4888]: I1124 00:46:20.541262 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:20 crc kubenswrapper[4888]: I1124 00:46:20.582488 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" podStartSLOduration=4.582444414 podStartE2EDuration="4.582444414s" podCreationTimestamp="2025-11-24 00:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:20.567212718 +0000 UTC m=+1283.149896762" watchObservedRunningTime="2025-11-24 00:46:20.582444414 +0000 UTC m=+1283.165128458" Nov 24 00:46:20 crc kubenswrapper[4888]: I1124 00:46:20.628352 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 00:46:21 crc kubenswrapper[4888]: I1124 00:46:21.562653 4888 generic.go:334] "Generic (PLEG): container finished" podID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerID="9a537b028d2d3cec56d995f0ceae9e459ebdd004cc57721653b0a43d0a0f6fcc" exitCode=0 Nov 24 00:46:21 crc kubenswrapper[4888]: I1124 00:46:21.562740 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerDied","Data":"9a537b028d2d3cec56d995f0ceae9e459ebdd004cc57721653b0a43d0a0f6fcc"} Nov 24 00:46:21 crc kubenswrapper[4888]: I1124 00:46:21.917084 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106040 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106390 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106549 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106573 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106626 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6gfm\" (UniqueName: \"kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106661 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.106699 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd\") pod \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\" (UID: \"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d\") " Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.107516 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.107622 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.115961 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts" (OuterVolumeSpecName: "scripts") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.117519 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm" (OuterVolumeSpecName: "kube-api-access-h6gfm") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "kube-api-access-h6gfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.146992 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.212672 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.212714 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.212726 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.212737 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.212748 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6gfm\" (UniqueName: \"kubernetes.io/projected/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-kube-api-access-h6gfm\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.223040 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.235339 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data" (OuterVolumeSpecName: "config-data") pod "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" (UID: "dafb43c7-844e-46f0-b94c-8c3aac9f0b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.314552 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.314583 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.578009 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dafb43c7-844e-46f0-b94c-8c3aac9f0b6d","Type":"ContainerDied","Data":"3307f98ef9203eea06fc6bf598cd42030c8ec51c60be7376c1871a135980e208"} Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.578336 4888 scope.go:117] "RemoveContainer" containerID="0aec7dd6106bfa49d067173e07a238ba214d214103b905dd9d140dc221ba581a" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.578469 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.583364 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-559fcf88cb-n5qbg" event={"ID":"eb2f6347-c0a7-4af3-b9f2-7df93d807a88","Type":"ContainerStarted","Data":"2f871711fc58e2e3ccc4faf37902885ca1bd1fa97007a859cbcdb23fd8c2d815"} Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.584137 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.597942 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" event={"ID":"a481e8b8-30cb-4449-99fe-24929099c3b5","Type":"ContainerStarted","Data":"d9026948f54728371553e66f310bb0495486e1ee0edd229324d0e3a7607bffc8"} Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.597988 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.619556 4888 scope.go:117] "RemoveContainer" containerID="7f7c9697d077cf51363b1e8705d28aa75042e3a8c5485bb4835f0df4f74825f6" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.628123 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-559fcf88cb-n5qbg" podStartSLOduration=3.307119799 podStartE2EDuration="6.628102655s" podCreationTimestamp="2025-11-24 00:46:16 +0000 UTC" firstStartedPulling="2025-11-24 00:46:18.300783855 +0000 UTC m=+1280.883467899" lastFinishedPulling="2025-11-24 00:46:21.621766711 +0000 UTC m=+1284.204450755" observedRunningTime="2025-11-24 00:46:22.622061026 +0000 UTC m=+1285.204745070" watchObservedRunningTime="2025-11-24 00:46:22.628102655 +0000 UTC m=+1285.210786699" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.658468 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" podStartSLOduration=3.187437291 podStartE2EDuration="6.658449153s" podCreationTimestamp="2025-11-24 00:46:16 +0000 UTC" firstStartedPulling="2025-11-24 00:46:18.14721574 +0000 UTC m=+1280.729899784" lastFinishedPulling="2025-11-24 00:46:21.618227492 +0000 UTC m=+1284.200911646" observedRunningTime="2025-11-24 00:46:22.647337833 +0000 UTC m=+1285.230021877" watchObservedRunningTime="2025-11-24 00:46:22.658449153 +0000 UTC m=+1285.241133197" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.664324 4888 scope.go:117] "RemoveContainer" containerID="9a537b028d2d3cec56d995f0ceae9e459ebdd004cc57721653b0a43d0a0f6fcc" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.694070 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.708259 4888 scope.go:117] "RemoveContainer" containerID="a3afaf059386d73e7c4f1ee441f9758e7849ba240a360d9ece19c1c7045a1a7b" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.709429 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.757882 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:22 crc kubenswrapper[4888]: E1124 00:46:22.758609 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="proxy-httpd" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.758628 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="proxy-httpd" Nov 24 00:46:22 crc kubenswrapper[4888]: E1124 00:46:22.758672 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-notification-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.758679 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-notification-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: E1124 00:46:22.758699 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="sg-core" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.779615 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="sg-core" Nov 24 00:46:22 crc kubenswrapper[4888]: E1124 00:46:22.779697 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-central-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.779712 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-central-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.780629 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="sg-core" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.780661 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-notification-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.780711 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="ceilometer-central-agent" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.780733 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" containerName="proxy-httpd" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.785307 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.792545 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.792758 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.845896 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938718 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938755 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938785 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crjr5\" (UniqueName: \"kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938855 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938874 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938895 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:22 crc kubenswrapper[4888]: I1124 00:46:22.938955 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.040997 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041099 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041121 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041259 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crjr5\" (UniqueName: \"kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041972 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041538 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.041999 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.042080 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.042907 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.049519 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.050695 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.051418 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.079866 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.120174 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crjr5\" (UniqueName: \"kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5\") pod \"ceilometer-0\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.151507 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.594310 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.594607 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.594659 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.595374 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.595417 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c" gracePeriod=600 Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.740887 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.742306 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.759682 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.763446 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.780391 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.788067 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.801551 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.803011 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.825736 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.851282 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.864917 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.864966 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkwmz\" (UniqueName: \"kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865026 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865062 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb742\" (UniqueName: \"kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865089 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865121 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865139 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.865222 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966587 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966640 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb742\" (UniqueName: \"kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966706 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966751 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966769 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966893 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966918 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966938 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966961 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkwmz\" (UniqueName: \"kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.966996 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqk9t\" (UniqueName: \"kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.974760 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.974842 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.975240 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.984750 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.984874 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.985561 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:23 crc kubenswrapper[4888]: I1124 00:46:23.986672 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb742\" (UniqueName: \"kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742\") pod \"heat-engine-595c95d7db-65z8t\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.009182 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkwmz\" (UniqueName: \"kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz\") pod \"heat-cfnapi-755f4964f5-4hwcx\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.069413 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.069503 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.069635 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.069699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqk9t\" (UniqueName: \"kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.073559 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.081513 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.083783 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.088368 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqk9t\" (UniqueName: \"kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t\") pod \"heat-api-8675c5c6d8-qlfv8\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.116434 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.188053 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.265038 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dafb43c7-844e-46f0-b94c-8c3aac9f0b6d" path="/var/lib/kubelet/pods/dafb43c7-844e-46f0-b94c-8c3aac9f0b6d/volumes" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.285175 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.668825 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.700495 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c" exitCode=0 Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.700631 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c"} Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.700713 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153"} Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.700776 4888 scope.go:117] "RemoveContainer" containerID="ac95392c2ff3024531af09e93551e6d9f56134f5a95f11b9ccc72fce2a4984c5" Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.707559 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerStarted","Data":"eb914d806bf48b50c157a36e9dc417e805b583b93d2b9e20e7b4f2e323ca0c91"} Nov 24 00:46:24 crc kubenswrapper[4888]: I1124 00:46:24.831480 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.152431 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:25 crc kubenswrapper[4888]: W1124 00:46:25.185938 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a0af49_0cd1_44d2_9b0c_438b5762db37.slice/crio-3fea65d497a7909fd9da4c6943aea71ee74d014712feec0a321c0645594f86bd WatchSource:0}: Error finding container 3fea65d497a7909fd9da4c6943aea71ee74d014712feec0a321c0645594f86bd: Status 404 returned error can't find the container with id 3fea65d497a7909fd9da4c6943aea71ee74d014712feec0a321c0645594f86bd Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.724195 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerStarted","Data":"f2768dfe2491c4ec4446cac7192172c2bf604e1b7a18c48b3356732a36a6f891"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.724543 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerStarted","Data":"3fea65d497a7909fd9da4c6943aea71ee74d014712feec0a321c0645594f86bd"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.724586 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.729198 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-595c95d7db-65z8t" event={"ID":"6865e42c-53d9-44fa-8fb2-77e9201fc53f","Type":"ContainerStarted","Data":"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.729379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-595c95d7db-65z8t" event={"ID":"6865e42c-53d9-44fa-8fb2-77e9201fc53f","Type":"ContainerStarted","Data":"36aa6b627a9e43e8f76d68913988e887e635f5c4dae335877d044e5f5aed7d9b"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.730408 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.734282 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerStarted","Data":"d493449e3cb1b2683243a395786d1f258ab3421259c72b1de732c4e20acca946"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.734323 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerStarted","Data":"f8246ad0d55f9629d3c4dcdcb3a401efa8a093640e8198c1d4c99e77bfe4e3a6"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.738100 4888 generic.go:334] "Generic (PLEG): container finished" podID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerID="1e229ed340f811f750dc1c03d5ec52a74ec6ec12a878e564adb7d7da461bf291" exitCode=1 Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.738141 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" event={"ID":"d2bab3ad-9998-4e24-8219-5bc6094a84f7","Type":"ContainerDied","Data":"1e229ed340f811f750dc1c03d5ec52a74ec6ec12a878e564adb7d7da461bf291"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.738168 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" event={"ID":"d2bab3ad-9998-4e24-8219-5bc6094a84f7","Type":"ContainerStarted","Data":"3ef38012ee9786230539ccf621cac12447b55a0016ac9999defa5c5f76af1bcc"} Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.740488 4888 scope.go:117] "RemoveContainer" containerID="1e229ed340f811f750dc1c03d5ec52a74ec6ec12a878e564adb7d7da461bf291" Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.756193 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-8675c5c6d8-qlfv8" podStartSLOduration=2.756175925 podStartE2EDuration="2.756175925s" podCreationTimestamp="2025-11-24 00:46:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:25.74847999 +0000 UTC m=+1288.331164034" watchObservedRunningTime="2025-11-24 00:46:25.756175925 +0000 UTC m=+1288.338859969" Nov 24 00:46:25 crc kubenswrapper[4888]: I1124 00:46:25.816431 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-595c95d7db-65z8t" podStartSLOduration=2.81641456 podStartE2EDuration="2.81641456s" podCreationTimestamp="2025-11-24 00:46:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:25.809330822 +0000 UTC m=+1288.392014866" watchObservedRunningTime="2025-11-24 00:46:25.81641456 +0000 UTC m=+1288.399098604" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.012033 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.754429 4888 generic.go:334] "Generic (PLEG): container finished" podID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" exitCode=1 Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.755596 4888 scope.go:117] "RemoveContainer" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.755597 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" event={"ID":"d2bab3ad-9998-4e24-8219-5bc6094a84f7","Type":"ContainerDied","Data":"3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43"} Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.756343 4888 scope.go:117] "RemoveContainer" containerID="1e229ed340f811f750dc1c03d5ec52a74ec6ec12a878e564adb7d7da461bf291" Nov 24 00:46:26 crc kubenswrapper[4888]: E1124 00:46:26.757059 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-755f4964f5-4hwcx_openstack(d2bab3ad-9998-4e24-8219-5bc6094a84f7)\"" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.758189 4888 generic.go:334] "Generic (PLEG): container finished" podID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerID="f2768dfe2491c4ec4446cac7192172c2bf604e1b7a18c48b3356732a36a6f891" exitCode=1 Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.758252 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerDied","Data":"f2768dfe2491c4ec4446cac7192172c2bf604e1b7a18c48b3356732a36a6f891"} Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.758946 4888 scope.go:117] "RemoveContainer" containerID="f2768dfe2491c4ec4446cac7192172c2bf604e1b7a18c48b3356732a36a6f891" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.767690 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerStarted","Data":"b5d019ebfdb50a6ef25065b787a2cc0b51dc219d87c2cc7ac4641707bb9a53dc"} Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.839935 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.917778 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:26 crc kubenswrapper[4888]: I1124 00:46:26.918009 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="dnsmasq-dns" containerID="cri-o://35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a" gracePeriod=10 Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.652767 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.676397 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.699245 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6dd96dd9bc-c9x6n" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793642 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793691 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793750 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793829 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n2c5\" (UniqueName: \"kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793859 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.793966 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0\") pod \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\" (UID: \"6cb420c6-5c18-4cde-9e34-5dec80c3b72c\") " Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.832254 4888 generic.go:334] "Generic (PLEG): container finished" podID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerID="35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a" exitCode=0 Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.832383 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" event={"ID":"6cb420c6-5c18-4cde-9e34-5dec80c3b72c","Type":"ContainerDied","Data":"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a"} Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.832418 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" event={"ID":"6cb420c6-5c18-4cde-9e34-5dec80c3b72c","Type":"ContainerDied","Data":"095cfe9239768b456acc4663f22e782d77307d38db7c661500b3335a1a8ba4a6"} Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.832475 4888 scope.go:117] "RemoveContainer" containerID="35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.832709 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-clvbv" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.879066 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5" (OuterVolumeSpecName: "kube-api-access-4n2c5") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "kube-api-access-4n2c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.901322 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n2c5\" (UniqueName: \"kubernetes.io/projected/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-kube-api-access-4n2c5\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.919322 4888 scope.go:117] "RemoveContainer" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" Nov 24 00:46:27 crc kubenswrapper[4888]: E1124 00:46:27.920600 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-755f4964f5-4hwcx_openstack(d2bab3ad-9998-4e24-8219-5bc6094a84f7)\"" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.949700 4888 scope.go:117] "RemoveContainer" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" Nov 24 00:46:27 crc kubenswrapper[4888]: E1124 00:46:27.950210 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8675c5c6d8-qlfv8_openstack(94a0af49-0cd1-44d2-9b0c-438b5762db37)\"" pod="openstack/heat-api-8675c5c6d8-qlfv8" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.950521 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerStarted","Data":"8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727"} Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.959435 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.959728 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" containerID="cri-o://2f871711fc58e2e3ccc4faf37902885ca1bd1fa97007a859cbcdb23fd8c2d815" gracePeriod=60 Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.986001 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.196:8004/healthcheck\": EOF" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.987256 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.196:8004/healthcheck\": EOF" Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.993458 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:27 crc kubenswrapper[4888]: I1124 00:46:27.993755 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerName="heat-cfnapi" containerID="cri-o://d9026948f54728371553e66f310bb0495486e1ee0edd229324d0e3a7607bffc8" gracePeriod=60 Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.058078 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:46:28 crc kubenswrapper[4888]: E1124 00:46:28.058618 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="init" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.058634 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="init" Nov 24 00:46:28 crc kubenswrapper[4888]: E1124 00:46:28.058649 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="dnsmasq-dns" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.058657 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="dnsmasq-dns" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.058936 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" containerName="dnsmasq-dns" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.074689 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.089869 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.090115 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.090182 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.093992 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.097358 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.097503 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.107730 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config" (OuterVolumeSpecName: "config") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.108157 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.119806 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.119974 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.131845 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.136061 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.142517 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.149783 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cb420c6-5c18-4cde-9e34-5dec80c3b72c" (UID: "6cb420c6-5c18-4cde-9e34-5dec80c3b72c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.210966 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.213536 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.213682 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.213839 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.213983 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.214062 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.214146 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkd4d\" (UniqueName: \"kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.214252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.214448 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.215395 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.215535 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkkpr\" (UniqueName: \"kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.215749 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.217724 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.217953 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.218045 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.218244 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cb420c6-5c18-4cde-9e34-5dec80c3b72c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.237137 4888 scope.go:117] "RemoveContainer" containerID="145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.277819 4888 scope.go:117] "RemoveContainer" containerID="35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a" Nov 24 00:46:28 crc kubenswrapper[4888]: E1124 00:46:28.279393 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a\": container with ID starting with 35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a not found: ID does not exist" containerID="35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.279430 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a"} err="failed to get container status \"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a\": rpc error: code = NotFound desc = could not find container \"35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a\": container with ID starting with 35fd78cabeacd1625750f9932310117f2ba026b161265a34142762e498c1799a not found: ID does not exist" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.279461 4888 scope.go:117] "RemoveContainer" containerID="145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574" Nov 24 00:46:28 crc kubenswrapper[4888]: E1124 00:46:28.281431 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574\": container with ID starting with 145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574 not found: ID does not exist" containerID="145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.281491 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574"} err="failed to get container status \"145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574\": rpc error: code = NotFound desc = could not find container \"145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574\": container with ID starting with 145036cd3024c1ab9c36a19964d9958c326d33db8ba9479ad79ae213b2cdf574 not found: ID does not exist" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.319958 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320019 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkkpr\" (UniqueName: \"kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320072 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320091 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320155 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320229 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320258 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320285 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320306 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320330 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkd4d\" (UniqueName: \"kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320357 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.320387 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.324897 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.325871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.327663 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.331712 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.334748 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.335419 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.336059 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.341802 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.342004 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.343828 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkd4d\" (UniqueName: \"kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d\") pod \"heat-cfnapi-7d4c46565b-l2868\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.344347 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.348925 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkkpr\" (UniqueName: \"kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr\") pod \"heat-api-69d5b6b95b-jhbt6\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.453009 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.466739 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-clvbv"] Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.469874 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.195:8000/healthcheck\": read tcp 10.217.0.2:59878->10.217.0.195:8000: read: connection reset by peer" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.556042 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:28 crc kubenswrapper[4888]: I1124 00:46:28.572654 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.005980 4888 generic.go:334] "Generic (PLEG): container finished" podID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" exitCode=1 Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.006450 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerDied","Data":"8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727"} Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.006483 4888 scope.go:117] "RemoveContainer" containerID="f2768dfe2491c4ec4446cac7192172c2bf604e1b7a18c48b3356732a36a6f891" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.006991 4888 scope.go:117] "RemoveContainer" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" Nov 24 00:46:29 crc kubenswrapper[4888]: E1124 00:46:29.007288 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8675c5c6d8-qlfv8_openstack(94a0af49-0cd1-44d2-9b0c-438b5762db37)\"" pod="openstack/heat-api-8675c5c6d8-qlfv8" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.038618 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerStarted","Data":"b0f371dac1534016e8e90105dffc91766a5cb206426d63e7c95e116b430e0dc3"} Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.039095 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.058983 4888 generic.go:334] "Generic (PLEG): container finished" podID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerID="d9026948f54728371553e66f310bb0495486e1ee0edd229324d0e3a7607bffc8" exitCode=0 Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.059051 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" event={"ID":"a481e8b8-30cb-4449-99fe-24929099c3b5","Type":"ContainerDied","Data":"d9026948f54728371553e66f310bb0495486e1ee0edd229324d0e3a7607bffc8"} Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.065768 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.279784142 podStartE2EDuration="7.065754942s" podCreationTimestamp="2025-11-24 00:46:22 +0000 UTC" firstStartedPulling="2025-11-24 00:46:23.863585526 +0000 UTC m=+1286.446269570" lastFinishedPulling="2025-11-24 00:46:27.649556326 +0000 UTC m=+1290.232240370" observedRunningTime="2025-11-24 00:46:29.059555129 +0000 UTC m=+1291.642239173" watchObservedRunningTime="2025-11-24 00:46:29.065754942 +0000 UTC m=+1291.648438986" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.120240 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.121122 4888 scope.go:117] "RemoveContainer" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" Nov 24 00:46:29 crc kubenswrapper[4888]: E1124 00:46:29.121389 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-755f4964f5-4hwcx_openstack(d2bab3ad-9998-4e24-8219-5bc6094a84f7)\"" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.121703 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.148307 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.233576 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.242741 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle\") pod \"a481e8b8-30cb-4449-99fe-24929099c3b5\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.242873 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c2lg\" (UniqueName: \"kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg\") pod \"a481e8b8-30cb-4449-99fe-24929099c3b5\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.242911 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data\") pod \"a481e8b8-30cb-4449-99fe-24929099c3b5\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.242960 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom\") pod \"a481e8b8-30cb-4449-99fe-24929099c3b5\" (UID: \"a481e8b8-30cb-4449-99fe-24929099c3b5\") " Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.247232 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.253949 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg" (OuterVolumeSpecName: "kube-api-access-5c2lg") pod "a481e8b8-30cb-4449-99fe-24929099c3b5" (UID: "a481e8b8-30cb-4449-99fe-24929099c3b5"). InnerVolumeSpecName "kube-api-access-5c2lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.255096 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a481e8b8-30cb-4449-99fe-24929099c3b5" (UID: "a481e8b8-30cb-4449-99fe-24929099c3b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.285795 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.285892 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.321707 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a481e8b8-30cb-4449-99fe-24929099c3b5" (UID: "a481e8b8-30cb-4449-99fe-24929099c3b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.346127 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.346181 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c2lg\" (UniqueName: \"kubernetes.io/projected/a481e8b8-30cb-4449-99fe-24929099c3b5-kube-api-access-5c2lg\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.346196 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.371664 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data" (OuterVolumeSpecName: "config-data") pod "a481e8b8-30cb-4449-99fe-24929099c3b5" (UID: "a481e8b8-30cb-4449-99fe-24929099c3b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:29 crc kubenswrapper[4888]: I1124 00:46:29.448322 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a481e8b8-30cb-4449-99fe-24929099c3b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.091543 4888 scope.go:117] "RemoveContainer" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" Nov 24 00:46:30 crc kubenswrapper[4888]: E1124 00:46:30.092069 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8675c5c6d8-qlfv8_openstack(94a0af49-0cd1-44d2-9b0c-438b5762db37)\"" pod="openstack/heat-api-8675c5c6d8-qlfv8" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.119385 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4c46565b-l2868" event={"ID":"6f9f41a6-bbce-4876-b1f5-b955b1c19786","Type":"ContainerStarted","Data":"be619cd7214e54b2bf26e2c69c27ef5c7a16d950bedf4b4d10469236963ae4e8"} Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.119422 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4c46565b-l2868" event={"ID":"6f9f41a6-bbce-4876-b1f5-b955b1c19786","Type":"ContainerStarted","Data":"85f9403ee321b09862f2ea58a09d13dcf8e47a7d0d3dac754bcb459b2936a8a7"} Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.120570 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.130247 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69d5b6b95b-jhbt6" event={"ID":"9f2c7353-a414-429f-a6a6-f42ed6120ccf","Type":"ContainerStarted","Data":"e3b1c60baf33e1c69193901f5b70d3bd167c112a9db60773ac0f39f047c8f838"} Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.130288 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69d5b6b95b-jhbt6" event={"ID":"9f2c7353-a414-429f-a6a6-f42ed6120ccf","Type":"ContainerStarted","Data":"62fae4d3a1d25684384949176eec3a53f42fcd0883426470dd1e95a747c349e0"} Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.131070 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.133283 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.141902 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c4dff5d5b-9qwfl" event={"ID":"a481e8b8-30cb-4449-99fe-24929099c3b5","Type":"ContainerDied","Data":"3e30ebe604ce6d231b0fef04941f9d694558f32abf5b01ad9f9f00f56b621327"} Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.141944 4888 scope.go:117] "RemoveContainer" containerID="d9026948f54728371553e66f310bb0495486e1ee0edd229324d0e3a7607bffc8" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.143259 4888 scope.go:117] "RemoveContainer" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" Nov 24 00:46:30 crc kubenswrapper[4888]: E1124 00:46:30.143457 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-755f4964f5-4hwcx_openstack(d2bab3ad-9998-4e24-8219-5bc6094a84f7)\"" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.154500 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7d4c46565b-l2868" podStartSLOduration=3.15448418 podStartE2EDuration="3.15448418s" podCreationTimestamp="2025-11-24 00:46:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:30.144261924 +0000 UTC m=+1292.726945968" watchObservedRunningTime="2025-11-24 00:46:30.15448418 +0000 UTC m=+1292.737168224" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.213250 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69d5b6b95b-jhbt6" podStartSLOduration=3.213229183 podStartE2EDuration="3.213229183s" podCreationTimestamp="2025-11-24 00:46:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:46:30.205211319 +0000 UTC m=+1292.787895363" watchObservedRunningTime="2025-11-24 00:46:30.213229183 +0000 UTC m=+1292.795913227" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.295555 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cb420c6-5c18-4cde-9e34-5dec80c3b72c" path="/var/lib/kubelet/pods/6cb420c6-5c18-4cde-9e34-5dec80c3b72c/volumes" Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.306594 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:30 crc kubenswrapper[4888]: I1124 00:46:30.310162 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7c4dff5d5b-9qwfl"] Nov 24 00:46:31 crc kubenswrapper[4888]: I1124 00:46:31.146549 4888 scope.go:117] "RemoveContainer" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" Nov 24 00:46:31 crc kubenswrapper[4888]: E1124 00:46:31.147147 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8675c5c6d8-qlfv8_openstack(94a0af49-0cd1-44d2-9b0c-438b5762db37)\"" pod="openstack/heat-api-8675c5c6d8-qlfv8" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.257226 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" path="/var/lib/kubelet/pods/a481e8b8-30cb-4449-99fe-24929099c3b5/volumes" Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.434242 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.434974 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-central-agent" containerID="cri-o://f8246ad0d55f9629d3c4dcdcb3a401efa8a093640e8198c1d4c99e77bfe4e3a6" gracePeriod=30 Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.435230 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-notification-agent" containerID="cri-o://d493449e3cb1b2683243a395786d1f258ab3421259c72b1de732c4e20acca946" gracePeriod=30 Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.435349 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="proxy-httpd" containerID="cri-o://b0f371dac1534016e8e90105dffc91766a5cb206426d63e7c95e116b430e0dc3" gracePeriod=30 Nov 24 00:46:32 crc kubenswrapper[4888]: I1124 00:46:32.435177 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="sg-core" containerID="cri-o://b5d019ebfdb50a6ef25065b787a2cc0b51dc219d87c2cc7ac4641707bb9a53dc" gracePeriod=30 Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.175769 4888 generic.go:334] "Generic (PLEG): container finished" podID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerID="b0f371dac1534016e8e90105dffc91766a5cb206426d63e7c95e116b430e0dc3" exitCode=0 Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.176054 4888 generic.go:334] "Generic (PLEG): container finished" podID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerID="b5d019ebfdb50a6ef25065b787a2cc0b51dc219d87c2cc7ac4641707bb9a53dc" exitCode=2 Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.176066 4888 generic.go:334] "Generic (PLEG): container finished" podID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerID="d493449e3cb1b2683243a395786d1f258ab3421259c72b1de732c4e20acca946" exitCode=0 Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.175840 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerDied","Data":"b0f371dac1534016e8e90105dffc91766a5cb206426d63e7c95e116b430e0dc3"} Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.176101 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerDied","Data":"b5d019ebfdb50a6ef25065b787a2cc0b51dc219d87c2cc7ac4641707bb9a53dc"} Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.176116 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerDied","Data":"d493449e3cb1b2683243a395786d1f258ab3421259c72b1de732c4e20acca946"} Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.379839 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.196:8004/healthcheck\": read tcp 10.217.0.2:55510->10.217.0.196:8004: read: connection reset by peer" Nov 24 00:46:33 crc kubenswrapper[4888]: I1124 00:46:33.380304 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.196:8004/healthcheck\": dial tcp 10.217.0.196:8004: connect: connection refused" Nov 24 00:46:34 crc kubenswrapper[4888]: I1124 00:46:34.191178 4888 generic.go:334] "Generic (PLEG): container finished" podID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerID="2f871711fc58e2e3ccc4faf37902885ca1bd1fa97007a859cbcdb23fd8c2d815" exitCode=0 Nov 24 00:46:34 crc kubenswrapper[4888]: I1124 00:46:34.191370 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-559fcf88cb-n5qbg" event={"ID":"eb2f6347-c0a7-4af3-b9f2-7df93d807a88","Type":"ContainerDied","Data":"2f871711fc58e2e3ccc4faf37902885ca1bd1fa97007a859cbcdb23fd8c2d815"} Nov 24 00:46:35 crc kubenswrapper[4888]: I1124 00:46:35.205236 4888 generic.go:334] "Generic (PLEG): container finished" podID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerID="f8246ad0d55f9629d3c4dcdcb3a401efa8a093640e8198c1d4c99e77bfe4e3a6" exitCode=0 Nov 24 00:46:35 crc kubenswrapper[4888]: I1124 00:46:35.205394 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerDied","Data":"f8246ad0d55f9629d3c4dcdcb3a401efa8a093640e8198c1d4c99e77bfe4e3a6"} Nov 24 00:46:36 crc kubenswrapper[4888]: I1124 00:46:36.780443 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.137853 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.228614 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a036e658-146d-4eb0-a197-939e6bbc31f8","Type":"ContainerStarted","Data":"51c1615579878a1e3a04721be70bc342037c820b9dc13e2d278709e4a8ed8bab"} Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.231506 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-559fcf88cb-n5qbg" event={"ID":"eb2f6347-c0a7-4af3-b9f2-7df93d807a88","Type":"ContainerDied","Data":"a5dfec0fa404e06253ab457943f0f32363b2b4fb222c615b305691e5a126c5c5"} Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.231564 4888 scope.go:117] "RemoveContainer" containerID="2f871711fc58e2e3ccc4faf37902885ca1bd1fa97007a859cbcdb23fd8c2d815" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.231724 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-559fcf88cb-n5qbg" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.250554 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle\") pod \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.250637 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdx76\" (UniqueName: \"kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76\") pod \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.250663 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom\") pod \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.250687 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data\") pod \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\" (UID: \"eb2f6347-c0a7-4af3-b9f2-7df93d807a88\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.255772 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76" (OuterVolumeSpecName: "kube-api-access-rdx76") pod "eb2f6347-c0a7-4af3-b9f2-7df93d807a88" (UID: "eb2f6347-c0a7-4af3-b9f2-7df93d807a88"). InnerVolumeSpecName "kube-api-access-rdx76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.257103 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eb2f6347-c0a7-4af3-b9f2-7df93d807a88" (UID: "eb2f6347-c0a7-4af3-b9f2-7df93d807a88"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.269840 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.095726048 podStartE2EDuration="20.269821111s" podCreationTimestamp="2025-11-24 00:46:17 +0000 UTC" firstStartedPulling="2025-11-24 00:46:18.697560612 +0000 UTC m=+1281.280244656" lastFinishedPulling="2025-11-24 00:46:36.871655675 +0000 UTC m=+1299.454339719" observedRunningTime="2025-11-24 00:46:37.248954777 +0000 UTC m=+1299.831638821" watchObservedRunningTime="2025-11-24 00:46:37.269821111 +0000 UTC m=+1299.852505155" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.279368 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.303160 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb2f6347-c0a7-4af3-b9f2-7df93d807a88" (UID: "eb2f6347-c0a7-4af3-b9f2-7df93d807a88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.320092 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data" (OuterVolumeSpecName: "config-data") pod "eb2f6347-c0a7-4af3-b9f2-7df93d807a88" (UID: "eb2f6347-c0a7-4af3-b9f2-7df93d807a88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352337 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352425 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352497 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352512 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352568 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352610 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.352654 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crjr5\" (UniqueName: \"kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5\") pod \"01e897dc-04ee-4534-9106-9dfd9cd49b27\" (UID: \"01e897dc-04ee-4534-9106-9dfd9cd49b27\") " Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353237 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353253 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdx76\" (UniqueName: \"kubernetes.io/projected/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-kube-api-access-rdx76\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353263 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353272 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2f6347-c0a7-4af3-b9f2-7df93d807a88-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353406 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.353751 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.356377 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts" (OuterVolumeSpecName: "scripts") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.358046 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5" (OuterVolumeSpecName: "kube-api-access-crjr5") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "kube-api-access-crjr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.391488 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.455273 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.455302 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.455311 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.455320 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01e897dc-04ee-4534-9106-9dfd9cd49b27-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.455328 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crjr5\" (UniqueName: \"kubernetes.io/projected/01e897dc-04ee-4534-9106-9dfd9cd49b27-kube-api-access-crjr5\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.468061 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data" (OuterVolumeSpecName: "config-data") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.482253 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01e897dc-04ee-4534-9106-9dfd9cd49b27" (UID: "01e897dc-04ee-4534-9106-9dfd9cd49b27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.559130 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.559360 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e897dc-04ee-4534-9106-9dfd9cd49b27-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.572043 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:37 crc kubenswrapper[4888]: I1124 00:46:37.584139 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-559fcf88cb-n5qbg"] Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.244007 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01e897dc-04ee-4534-9106-9dfd9cd49b27","Type":"ContainerDied","Data":"eb914d806bf48b50c157a36e9dc417e805b583b93d2b9e20e7b4f2e323ca0c91"} Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.244060 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.244105 4888 scope.go:117] "RemoveContainer" containerID="b0f371dac1534016e8e90105dffc91766a5cb206426d63e7c95e116b430e0dc3" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.259172 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" path="/var/lib/kubelet/pods/eb2f6347-c0a7-4af3-b9f2-7df93d807a88/volumes" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.272549 4888 scope.go:117] "RemoveContainer" containerID="b5d019ebfdb50a6ef25065b787a2cc0b51dc219d87c2cc7ac4641707bb9a53dc" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.298384 4888 scope.go:117] "RemoveContainer" containerID="d493449e3cb1b2683243a395786d1f258ab3421259c72b1de732c4e20acca946" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.298846 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.311337 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323017 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323534 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-central-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323556 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-central-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323574 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="sg-core" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323583 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="sg-core" Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323595 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="proxy-httpd" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323604 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="proxy-httpd" Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323627 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323635 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323652 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-notification-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323660 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-notification-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: E1124 00:46:38.323680 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerName="heat-cfnapi" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323688 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerName="heat-cfnapi" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323951 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-central-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323972 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="ceilometer-notification-agent" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323985 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.323995 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="proxy-httpd" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.324007 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" containerName="sg-core" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.324022 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a481e8b8-30cb-4449-99fe-24929099c3b5" containerName="heat-cfnapi" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.330591 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.334996 4888 scope.go:117] "RemoveContainer" containerID="f8246ad0d55f9629d3c4dcdcb3a401efa8a093640e8198c1d4c99e77bfe4e3a6" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.335447 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.337223 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.383044 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.485781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.485835 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.485863 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.486018 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.486084 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.486333 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.486501 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmfx2\" (UniqueName: \"kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.587791 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588120 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588234 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588241 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588484 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmfx2\" (UniqueName: \"kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588612 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.588641 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.589511 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.595878 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.599558 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.603932 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.609206 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.610090 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmfx2\" (UniqueName: \"kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2\") pod \"ceilometer-0\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " pod="openstack/ceilometer-0" Nov 24 00:46:38 crc kubenswrapper[4888]: I1124 00:46:38.662282 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:39 crc kubenswrapper[4888]: I1124 00:46:39.404925 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.257056 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e897dc-04ee-4534-9106-9dfd9cd49b27" path="/var/lib/kubelet/pods/01e897dc-04ee-4534-9106-9dfd9cd49b27/volumes" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.258294 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.269028 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerStarted","Data":"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59"} Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.269066 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerStarted","Data":"4b0fec4618ed9b34c61c1e33a0ed8cc5e4d6574d8304d835e94c313c6fdc2c39"} Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.318353 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.468410 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.543390 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.793344 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.946248 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle\") pod \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.946438 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data\") pod \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.946471 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom\") pod \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.946561 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkwmz\" (UniqueName: \"kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz\") pod \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\" (UID: \"d2bab3ad-9998-4e24-8219-5bc6094a84f7\") " Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.950628 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz" (OuterVolumeSpecName: "kube-api-access-lkwmz") pod "d2bab3ad-9998-4e24-8219-5bc6094a84f7" (UID: "d2bab3ad-9998-4e24-8219-5bc6094a84f7"). InnerVolumeSpecName "kube-api-access-lkwmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.952008 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d2bab3ad-9998-4e24-8219-5bc6094a84f7" (UID: "d2bab3ad-9998-4e24-8219-5bc6094a84f7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:40 crc kubenswrapper[4888]: I1124 00:46:40.991523 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2bab3ad-9998-4e24-8219-5bc6094a84f7" (UID: "d2bab3ad-9998-4e24-8219-5bc6094a84f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.013314 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data" (OuterVolumeSpecName: "config-data") pod "d2bab3ad-9998-4e24-8219-5bc6094a84f7" (UID: "d2bab3ad-9998-4e24-8219-5bc6094a84f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.050026 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkwmz\" (UniqueName: \"kubernetes.io/projected/d2bab3ad-9998-4e24-8219-5bc6094a84f7-kube-api-access-lkwmz\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.050443 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.050535 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.050611 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2bab3ad-9998-4e24-8219-5bc6094a84f7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.122616 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.256663 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqk9t\" (UniqueName: \"kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t\") pod \"94a0af49-0cd1-44d2-9b0c-438b5762db37\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.257323 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom\") pod \"94a0af49-0cd1-44d2-9b0c-438b5762db37\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.257482 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data\") pod \"94a0af49-0cd1-44d2-9b0c-438b5762db37\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.257644 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle\") pod \"94a0af49-0cd1-44d2-9b0c-438b5762db37\" (UID: \"94a0af49-0cd1-44d2-9b0c-438b5762db37\") " Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.262347 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "94a0af49-0cd1-44d2-9b0c-438b5762db37" (UID: "94a0af49-0cd1-44d2-9b0c-438b5762db37"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.271732 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t" (OuterVolumeSpecName: "kube-api-access-qqk9t") pod "94a0af49-0cd1-44d2-9b0c-438b5762db37" (UID: "94a0af49-0cd1-44d2-9b0c-438b5762db37"). InnerVolumeSpecName "kube-api-access-qqk9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.285469 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" event={"ID":"d2bab3ad-9998-4e24-8219-5bc6094a84f7","Type":"ContainerDied","Data":"3ef38012ee9786230539ccf621cac12447b55a0016ac9999defa5c5f76af1bcc"} Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.285546 4888 scope.go:117] "RemoveContainer" containerID="3d8f91c6c9e7f732dbe4e4f4601e125ad800339f637a7a215c6ff1ec144fef43" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.285549 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755f4964f5-4hwcx" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.288088 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8675c5c6d8-qlfv8" event={"ID":"94a0af49-0cd1-44d2-9b0c-438b5762db37","Type":"ContainerDied","Data":"3fea65d497a7909fd9da4c6943aea71ee74d014712feec0a321c0645594f86bd"} Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.288162 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8675c5c6d8-qlfv8" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.303795 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94a0af49-0cd1-44d2-9b0c-438b5762db37" (UID: "94a0af49-0cd1-44d2-9b0c-438b5762db37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.345907 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data" (OuterVolumeSpecName: "config-data") pod "94a0af49-0cd1-44d2-9b0c-438b5762db37" (UID: "94a0af49-0cd1-44d2-9b0c-438b5762db37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.362112 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqk9t\" (UniqueName: \"kubernetes.io/projected/94a0af49-0cd1-44d2-9b0c-438b5762db37-kube-api-access-qqk9t\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.362353 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.362420 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.362478 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a0af49-0cd1-44d2-9b0c-438b5762db37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.424263 4888 scope.go:117] "RemoveContainer" containerID="8a52383c4fcad647813fe529bf9b9cb72e8e0bd4e657cdb48d013a963c6ea727" Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.429322 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.441158 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-755f4964f5-4hwcx"] Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.621091 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:41 crc kubenswrapper[4888]: I1124 00:46:41.630871 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-8675c5c6d8-qlfv8"] Nov 24 00:46:42 crc kubenswrapper[4888]: I1124 00:46:42.259714 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" path="/var/lib/kubelet/pods/94a0af49-0cd1-44d2-9b0c-438b5762db37/volumes" Nov 24 00:46:42 crc kubenswrapper[4888]: I1124 00:46:42.260785 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" path="/var/lib/kubelet/pods/d2bab3ad-9998-4e24-8219-5bc6094a84f7/volumes" Nov 24 00:46:42 crc kubenswrapper[4888]: I1124 00:46:42.298867 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerStarted","Data":"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2"} Nov 24 00:46:43 crc kubenswrapper[4888]: I1124 00:46:43.316174 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerStarted","Data":"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617"} Nov 24 00:46:44 crc kubenswrapper[4888]: I1124 00:46:44.297078 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:46:44 crc kubenswrapper[4888]: I1124 00:46:44.408288 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:44 crc kubenswrapper[4888]: I1124 00:46:44.408558 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-79868646c7-2h6p7" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" containerName="heat-engine" containerID="cri-o://699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" gracePeriod=60 Nov 24 00:46:45 crc kubenswrapper[4888]: I1124 00:46:45.381659 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerStarted","Data":"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e"} Nov 24 00:46:45 crc kubenswrapper[4888]: I1124 00:46:45.382125 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:46:45 crc kubenswrapper[4888]: I1124 00:46:45.409098 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.693467357 podStartE2EDuration="7.409077196s" podCreationTimestamp="2025-11-24 00:46:38 +0000 UTC" firstStartedPulling="2025-11-24 00:46:39.427002799 +0000 UTC m=+1302.009686843" lastFinishedPulling="2025-11-24 00:46:44.142612638 +0000 UTC m=+1306.725296682" observedRunningTime="2025-11-24 00:46:45.400539117 +0000 UTC m=+1307.983223171" watchObservedRunningTime="2025-11-24 00:46:45.409077196 +0000 UTC m=+1307.991761240" Nov 24 00:46:45 crc kubenswrapper[4888]: I1124 00:46:45.947462 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289098 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-krvgq"] Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.289470 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289487 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.289505 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289511 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.289523 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289529 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.289539 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289544 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289730 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289745 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a0af49-0cd1-44d2-9b0c-438b5762db37" containerName="heat-api" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289756 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.289764 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2bab3ad-9998-4e24-8219-5bc6094a84f7" containerName="heat-cfnapi" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.290416 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.296065 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7nls\" (UniqueName: \"kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.296335 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.365479 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-krvgq"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.399376 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7nls\" (UniqueName: \"kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.399435 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.400180 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.438882 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-n7966"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.440544 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.459352 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7nls\" (UniqueName: \"kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls\") pod \"nova-api-db-create-krvgq\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.479445 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n7966"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.538880 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a6aa-account-create-25cm2"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.540831 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.545673 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.556370 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a6aa-account-create-25cm2"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.603524 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7kgw5"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.604775 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607006 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607241 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607332 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dglp\" (UniqueName: \"kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607405 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgpsh\" (UniqueName: \"kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607558 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.607598 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgw9n\" (UniqueName: \"kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.625681 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7kgw5"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.634863 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.708865 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709346 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f706-account-create-p74tv"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709386 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dglp\" (UniqueName: \"kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709428 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgpsh\" (UniqueName: \"kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709517 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709543 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgw9n\" (UniqueName: \"kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.709585 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.710242 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.710713 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.710765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.711631 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.714047 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.724548 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f706-account-create-p74tv"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.730422 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dglp\" (UniqueName: \"kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp\") pod \"nova-api-a6aa-account-create-25cm2\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.730754 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgw9n\" (UniqueName: \"kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n\") pod \"nova-cell0-db-create-n7966\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.732508 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgpsh\" (UniqueName: \"kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh\") pod \"nova-cell1-db-create-7kgw5\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.747638 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.748608 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.749458 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:46:46 crc kubenswrapper[4888]: E1124 00:46:46.749482 4888 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-79868646c7-2h6p7" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" containerName="heat-engine" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.819033 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4cv\" (UniqueName: \"kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.826028 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.836870 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.883271 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.927750 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8291-account-create-mpq4l"] Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.931030 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4cv\" (UniqueName: \"kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.931097 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.931662 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.931939 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.938060 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.955966 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.959092 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4cv\" (UniqueName: \"kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv\") pod \"nova-cell0-f706-account-create-p74tv\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:46 crc kubenswrapper[4888]: I1124 00:46:46.966874 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8291-account-create-mpq4l"] Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.033284 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k27gb\" (UniqueName: \"kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.033410 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.135319 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k27gb\" (UniqueName: \"kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.135589 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.136628 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.136744 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-559fcf88cb-n5qbg" podUID="eb2f6347-c0a7-4af3-b9f2-7df93d807a88" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.196:8004/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.150449 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.159628 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k27gb\" (UniqueName: \"kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb\") pod \"nova-cell1-8291-account-create-mpq4l\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.214130 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-krvgq"] Nov 24 00:46:47 crc kubenswrapper[4888]: W1124 00:46:47.231882 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fa347cb_768e_4e7b_ba47_42f178dec4ab.slice/crio-1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880 WatchSource:0}: Error finding container 1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880: Status 404 returned error can't find the container with id 1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880 Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.261138 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.425763 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n7966"] Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.453031 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-krvgq" event={"ID":"4fa347cb-768e-4e7b-ba47-42f178dec4ab","Type":"ContainerStarted","Data":"1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880"} Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.453265 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-central-agent" containerID="cri-o://8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59" gracePeriod=30 Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.453549 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="proxy-httpd" containerID="cri-o://db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e" gracePeriod=30 Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.453661 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-notification-agent" containerID="cri-o://e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2" gracePeriod=30 Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.453724 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="sg-core" containerID="cri-o://4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617" gracePeriod=30 Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.636366 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a6aa-account-create-25cm2"] Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.663230 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f706-account-create-p74tv"] Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.801191 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7kgw5"] Nov 24 00:46:47 crc kubenswrapper[4888]: I1124 00:46:47.917924 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8291-account-create-mpq4l"] Nov 24 00:46:47 crc kubenswrapper[4888]: W1124 00:46:47.939285 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59b6f3b1_832a_4340_92dd_55b7e2797d49.slice/crio-c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f WatchSource:0}: Error finding container c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f: Status 404 returned error can't find the container with id c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.464838 4888 generic.go:334] "Generic (PLEG): container finished" podID="fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" containerID="79326d578c816a9c3b9bd2072e82d6f94e02c722d967770f2dc437b004d4e021" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.464898 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n7966" event={"ID":"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21","Type":"ContainerDied","Data":"79326d578c816a9c3b9bd2072e82d6f94e02c722d967770f2dc437b004d4e021"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.464922 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n7966" event={"ID":"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21","Type":"ContainerStarted","Data":"f69431cd4a915a5959ae37955aa3b9e5e5c2c979b03306b2647b5c9ef5933076"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.472426 4888 generic.go:334] "Generic (PLEG): container finished" podID="f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" containerID="2acd0defe8157d1daac8851fd7575f316bebaf11336fc50a578fb04d7ab5f947" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.472509 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7kgw5" event={"ID":"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61","Type":"ContainerDied","Data":"2acd0defe8157d1daac8851fd7575f316bebaf11336fc50a578fb04d7ab5f947"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.472543 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7kgw5" event={"ID":"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61","Type":"ContainerStarted","Data":"52fca231da339664eba84ae03cb5b29761749ddf22434d587524717f8de1c55a"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481695 4888 generic.go:334] "Generic (PLEG): container finished" podID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerID="db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481728 4888 generic.go:334] "Generic (PLEG): container finished" podID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerID="4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617" exitCode=2 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481736 4888 generic.go:334] "Generic (PLEG): container finished" podID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerID="e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481802 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerDied","Data":"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481845 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerDied","Data":"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.481855 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerDied","Data":"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.485590 4888 generic.go:334] "Generic (PLEG): container finished" podID="df146d80-370a-4eac-afb5-35e747ed7007" containerID="024131d8514d8124a940492cb306d59b98c899da4cb6288834a4ff72f8343f26" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.485659 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f706-account-create-p74tv" event={"ID":"df146d80-370a-4eac-afb5-35e747ed7007","Type":"ContainerDied","Data":"024131d8514d8124a940492cb306d59b98c899da4cb6288834a4ff72f8343f26"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.485684 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f706-account-create-p74tv" event={"ID":"df146d80-370a-4eac-afb5-35e747ed7007","Type":"ContainerStarted","Data":"c346fac2c2404fdbe88d9942a73e3892bc60de5ab7b3ef56a7f20a6a2bf3a57d"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.489426 4888 generic.go:334] "Generic (PLEG): container finished" podID="74731cdc-25ca-44c5-a01b-ebe728556384" containerID="5742acb8bd64d8d25648c70f6547a157a1e55fcb6c81f42f029b325309b57aad" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.489515 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a6aa-account-create-25cm2" event={"ID":"74731cdc-25ca-44c5-a01b-ebe728556384","Type":"ContainerDied","Data":"5742acb8bd64d8d25648c70f6547a157a1e55fcb6c81f42f029b325309b57aad"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.489536 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a6aa-account-create-25cm2" event={"ID":"74731cdc-25ca-44c5-a01b-ebe728556384","Type":"ContainerStarted","Data":"c9b5bbcdd1401df7d67fdebdbb5917e5f9e9584faae14c3d33e745f5c9fef289"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.491138 4888 generic.go:334] "Generic (PLEG): container finished" podID="4fa347cb-768e-4e7b-ba47-42f178dec4ab" containerID="8aa4efcffd04b0341bb00d54dc7bf77d11fd47c9503b5aaac3cf56eaabaabdf5" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.491178 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-krvgq" event={"ID":"4fa347cb-768e-4e7b-ba47-42f178dec4ab","Type":"ContainerDied","Data":"8aa4efcffd04b0341bb00d54dc7bf77d11fd47c9503b5aaac3cf56eaabaabdf5"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.496454 4888 generic.go:334] "Generic (PLEG): container finished" podID="59b6f3b1-832a-4340-92dd-55b7e2797d49" containerID="a963745166ee4f2cb69db0b8cb83c9697de979b3e2e9637074edce4745a29b42" exitCode=0 Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.496510 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8291-account-create-mpq4l" event={"ID":"59b6f3b1-832a-4340-92dd-55b7e2797d49","Type":"ContainerDied","Data":"a963745166ee4f2cb69db0b8cb83c9697de979b3e2e9637074edce4745a29b42"} Nov 24 00:46:48 crc kubenswrapper[4888]: I1124 00:46:48.496538 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8291-account-create-mpq4l" event={"ID":"59b6f3b1-832a-4340-92dd-55b7e2797d49","Type":"ContainerStarted","Data":"c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.056802 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.139432 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dglp\" (UniqueName: \"kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp\") pod \"74731cdc-25ca-44c5-a01b-ebe728556384\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.139489 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts\") pod \"74731cdc-25ca-44c5-a01b-ebe728556384\" (UID: \"74731cdc-25ca-44c5-a01b-ebe728556384\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.140664 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74731cdc-25ca-44c5-a01b-ebe728556384" (UID: "74731cdc-25ca-44c5-a01b-ebe728556384"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.155971 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp" (OuterVolumeSpecName: "kube-api-access-7dglp") pod "74731cdc-25ca-44c5-a01b-ebe728556384" (UID: "74731cdc-25ca-44c5-a01b-ebe728556384"). InnerVolumeSpecName "kube-api-access-7dglp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.243292 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dglp\" (UniqueName: \"kubernetes.io/projected/74731cdc-25ca-44c5-a01b-ebe728556384-kube-api-access-7dglp\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.243323 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74731cdc-25ca-44c5-a01b-ebe728556384-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.457321 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.464703 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.470357 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.476209 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.488395 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.544803 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f706-account-create-p74tv" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.545011 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f706-account-create-p74tv" event={"ID":"df146d80-370a-4eac-afb5-35e747ed7007","Type":"ContainerDied","Data":"c346fac2c2404fdbe88d9942a73e3892bc60de5ab7b3ef56a7f20a6a2bf3a57d"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.545052 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c346fac2c2404fdbe88d9942a73e3892bc60de5ab7b3ef56a7f20a6a2bf3a57d" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550169 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgpsh\" (UniqueName: \"kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh\") pod \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550258 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts\") pod \"59b6f3b1-832a-4340-92dd-55b7e2797d49\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550352 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts\") pod \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550487 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k27gb\" (UniqueName: \"kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb\") pod \"59b6f3b1-832a-4340-92dd-55b7e2797d49\" (UID: \"59b6f3b1-832a-4340-92dd-55b7e2797d49\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550544 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts\") pod \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550574 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql4cv\" (UniqueName: \"kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv\") pod \"df146d80-370a-4eac-afb5-35e747ed7007\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550652 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgw9n\" (UniqueName: \"kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n\") pod \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\" (UID: \"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550674 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7nls\" (UniqueName: \"kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls\") pod \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\" (UID: \"4fa347cb-768e-4e7b-ba47-42f178dec4ab\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550702 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts\") pod \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\" (UID: \"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.550733 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts\") pod \"df146d80-370a-4eac-afb5-35e747ed7007\" (UID: \"df146d80-370a-4eac-afb5-35e747ed7007\") " Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.551912 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59b6f3b1-832a-4340-92dd-55b7e2797d49" (UID: "59b6f3b1-832a-4340-92dd-55b7e2797d49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.552004 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" (UID: "fa3acc5c-2554-4d9c-b515-b01a8f0fbd21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.552418 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a6aa-account-create-25cm2" event={"ID":"74731cdc-25ca-44c5-a01b-ebe728556384","Type":"ContainerDied","Data":"c9b5bbcdd1401df7d67fdebdbb5917e5f9e9584faae14c3d33e745f5c9fef289"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.552452 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4fa347cb-768e-4e7b-ba47-42f178dec4ab" (UID: "4fa347cb-768e-4e7b-ba47-42f178dec4ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.552457 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9b5bbcdd1401df7d67fdebdbb5917e5f9e9584faae14c3d33e745f5c9fef289" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.552511 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a6aa-account-create-25cm2" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554333 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df146d80-370a-4eac-afb5-35e747ed7007" (UID: "df146d80-370a-4eac-afb5-35e747ed7007"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554353 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" (UID: "f18c5664-ccfc-4ae6-a9c2-5a543cca5d61"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554824 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554845 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554854 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df146d80-370a-4eac-afb5-35e747ed7007-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554864 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59b6f3b1-832a-4340-92dd-55b7e2797d49-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.554874 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fa347cb-768e-4e7b-ba47-42f178dec4ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.557961 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n" (OuterVolumeSpecName: "kube-api-access-cgw9n") pod "fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" (UID: "fa3acc5c-2554-4d9c-b515-b01a8f0fbd21"). InnerVolumeSpecName "kube-api-access-cgw9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.558185 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh" (OuterVolumeSpecName: "kube-api-access-mgpsh") pod "f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" (UID: "f18c5664-ccfc-4ae6-a9c2-5a543cca5d61"). InnerVolumeSpecName "kube-api-access-mgpsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.559674 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls" (OuterVolumeSpecName: "kube-api-access-d7nls") pod "4fa347cb-768e-4e7b-ba47-42f178dec4ab" (UID: "4fa347cb-768e-4e7b-ba47-42f178dec4ab"). InnerVolumeSpecName "kube-api-access-d7nls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.561308 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-krvgq" event={"ID":"4fa347cb-768e-4e7b-ba47-42f178dec4ab","Type":"ContainerDied","Data":"1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.561342 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f7daad20d7ea13c03afd4d4f4bf2e32b9dd515279f51d55804caa86a4d66880" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.561401 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-krvgq" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.563009 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb" (OuterVolumeSpecName: "kube-api-access-k27gb") pod "59b6f3b1-832a-4340-92dd-55b7e2797d49" (UID: "59b6f3b1-832a-4340-92dd-55b7e2797d49"). InnerVolumeSpecName "kube-api-access-k27gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.568738 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8291-account-create-mpq4l" event={"ID":"59b6f3b1-832a-4340-92dd-55b7e2797d49","Type":"ContainerDied","Data":"c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.569078 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f9545f44a25dac46c568c6eedafe688510890b019ca8606f6e09f3af9dd64f" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.569134 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8291-account-create-mpq4l" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.586695 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n7966" event={"ID":"fa3acc5c-2554-4d9c-b515-b01a8f0fbd21","Type":"ContainerDied","Data":"f69431cd4a915a5959ae37955aa3b9e5e5c2c979b03306b2647b5c9ef5933076"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.586953 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f69431cd4a915a5959ae37955aa3b9e5e5c2c979b03306b2647b5c9ef5933076" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.587093 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n7966" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.589500 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7kgw5" event={"ID":"f18c5664-ccfc-4ae6-a9c2-5a543cca5d61","Type":"ContainerDied","Data":"52fca231da339664eba84ae03cb5b29761749ddf22434d587524717f8de1c55a"} Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.589537 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52fca231da339664eba84ae03cb5b29761749ddf22434d587524717f8de1c55a" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.589585 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7kgw5" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.596069 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv" (OuterVolumeSpecName: "kube-api-access-ql4cv") pod "df146d80-370a-4eac-afb5-35e747ed7007" (UID: "df146d80-370a-4eac-afb5-35e747ed7007"). InnerVolumeSpecName "kube-api-access-ql4cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.657140 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgw9n\" (UniqueName: \"kubernetes.io/projected/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21-kube-api-access-cgw9n\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.657176 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7nls\" (UniqueName: \"kubernetes.io/projected/4fa347cb-768e-4e7b-ba47-42f178dec4ab-kube-api-access-d7nls\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.657186 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgpsh\" (UniqueName: \"kubernetes.io/projected/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61-kube-api-access-mgpsh\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.657195 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k27gb\" (UniqueName: \"kubernetes.io/projected/59b6f3b1-832a-4340-92dd-55b7e2797d49-kube-api-access-k27gb\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:50 crc kubenswrapper[4888]: I1124 00:46:50.657204 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql4cv\" (UniqueName: \"kubernetes.io/projected/df146d80-370a-4eac-afb5-35e747ed7007-kube-api-access-ql4cv\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.072846 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.190069 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jndxh\" (UniqueName: \"kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh\") pod \"241b6970-6ac3-49b0-8b51-8f727991f324\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.190107 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom\") pod \"241b6970-6ac3-49b0-8b51-8f727991f324\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.190163 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle\") pod \"241b6970-6ac3-49b0-8b51-8f727991f324\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.190304 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data\") pod \"241b6970-6ac3-49b0-8b51-8f727991f324\" (UID: \"241b6970-6ac3-49b0-8b51-8f727991f324\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.199088 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "241b6970-6ac3-49b0-8b51-8f727991f324" (UID: "241b6970-6ac3-49b0-8b51-8f727991f324"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.200065 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh" (OuterVolumeSpecName: "kube-api-access-jndxh") pod "241b6970-6ac3-49b0-8b51-8f727991f324" (UID: "241b6970-6ac3-49b0-8b51-8f727991f324"). InnerVolumeSpecName "kube-api-access-jndxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.221514 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "241b6970-6ac3-49b0-8b51-8f727991f324" (UID: "241b6970-6ac3-49b0-8b51-8f727991f324"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.247976 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data" (OuterVolumeSpecName: "config-data") pod "241b6970-6ac3-49b0-8b51-8f727991f324" (UID: "241b6970-6ac3-49b0-8b51-8f727991f324"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.292694 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jndxh\" (UniqueName: \"kubernetes.io/projected/241b6970-6ac3-49b0-8b51-8f727991f324-kube-api-access-jndxh\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.292943 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.293003 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.293055 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241b6970-6ac3-49b0-8b51-8f727991f324-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.316005 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394411 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmfx2\" (UniqueName: \"kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394466 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394611 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394650 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394701 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394739 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.394793 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd\") pod \"acfa2cf3-151c-45de-8da2-a77724d90ee7\" (UID: \"acfa2cf3-151c-45de-8da2-a77724d90ee7\") " Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.395611 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.395689 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.399169 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2" (OuterVolumeSpecName: "kube-api-access-vmfx2") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "kube-api-access-vmfx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.399810 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts" (OuterVolumeSpecName: "scripts") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.427544 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.481476 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.498593 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmfx2\" (UniqueName: \"kubernetes.io/projected/acfa2cf3-151c-45de-8da2-a77724d90ee7-kube-api-access-vmfx2\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.498798 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.498901 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.498963 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.499028 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.499085 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acfa2cf3-151c-45de-8da2-a77724d90ee7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.506459 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data" (OuterVolumeSpecName: "config-data") pod "acfa2cf3-151c-45de-8da2-a77724d90ee7" (UID: "acfa2cf3-151c-45de-8da2-a77724d90ee7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.600386 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acfa2cf3-151c-45de-8da2-a77724d90ee7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.607804 4888 generic.go:334] "Generic (PLEG): container finished" podID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerID="8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59" exitCode=0 Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.607873 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerDied","Data":"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59"} Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.607961 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.608194 4888 scope.go:117] "RemoveContainer" containerID="db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.608179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"acfa2cf3-151c-45de-8da2-a77724d90ee7","Type":"ContainerDied","Data":"4b0fec4618ed9b34c61c1e33a0ed8cc5e4d6574d8304d835e94c313c6fdc2c39"} Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.610800 4888 generic.go:334] "Generic (PLEG): container finished" podID="241b6970-6ac3-49b0-8b51-8f727991f324" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" exitCode=0 Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.610874 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-79868646c7-2h6p7" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.610884 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-79868646c7-2h6p7" event={"ID":"241b6970-6ac3-49b0-8b51-8f727991f324","Type":"ContainerDied","Data":"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875"} Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.610909 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-79868646c7-2h6p7" event={"ID":"241b6970-6ac3-49b0-8b51-8f727991f324","Type":"ContainerDied","Data":"4dd985148283d4a5e153768edecedcb2ac5c81b750a7c97d49b5092a74d9b03f"} Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.630468 4888 scope.go:117] "RemoveContainer" containerID="4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.644303 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.656501 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-79868646c7-2h6p7"] Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.657849 4888 scope.go:117] "RemoveContainer" containerID="e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.677419 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.683747 4888 scope.go:117] "RemoveContainer" containerID="8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.701674 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.726061 4888 scope.go:117] "RemoveContainer" containerID="db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728048 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e\": container with ID starting with db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e not found: ID does not exist" containerID="db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728088 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e"} err="failed to get container status \"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e\": rpc error: code = NotFound desc = could not find container \"db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e\": container with ID starting with db42410d4d197955e11ebfef655c880c6db6aacb8dc5ad7ff2e75067a45d976e not found: ID does not exist" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728112 4888 scope.go:117] "RemoveContainer" containerID="4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728203 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728569 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617\": container with ID starting with 4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617 not found: ID does not exist" containerID="4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728592 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-central-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728603 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-central-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728605 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617"} err="failed to get container status \"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617\": rpc error: code = NotFound desc = could not find container \"4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617\": container with ID starting with 4ffd2532e64a7942f54c30abcde6d6814de85e7f3ac06b17c95b76867ced5617 not found: ID does not exist" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728625 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="proxy-httpd" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728629 4888 scope.go:117] "RemoveContainer" containerID="e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728632 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="proxy-httpd" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728716 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df146d80-370a-4eac-afb5-35e747ed7007" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728727 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="df146d80-370a-4eac-afb5-35e747ed7007" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728748 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74731cdc-25ca-44c5-a01b-ebe728556384" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728757 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="74731cdc-25ca-44c5-a01b-ebe728556384" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728768 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728776 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728788 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="sg-core" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728796 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="sg-core" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728842 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728867 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728874 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b6f3b1-832a-4340-92dd-55b7e2797d49" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728880 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b6f3b1-832a-4340-92dd-55b7e2797d49" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728899 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-notification-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728905 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-notification-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728907 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2\": container with ID starting with e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2 not found: ID does not exist" containerID="e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728925 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2"} err="failed to get container status \"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2\": rpc error: code = NotFound desc = could not find container \"e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2\": container with ID starting with e4ded366011720d29ad09af6040c026c0eecec8874b04efcd7306d008f0a5fb2 not found: ID does not exist" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728937 4888 scope.go:117] "RemoveContainer" containerID="8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728946 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" containerName="heat-engine" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728955 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" containerName="heat-engine" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.728967 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa347cb-768e-4e7b-ba47-42f178dec4ab" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.728973 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa347cb-768e-4e7b-ba47-42f178dec4ab" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.729126 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59\": container with ID starting with 8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59 not found: ID does not exist" containerID="8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729142 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59"} err="failed to get container status \"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59\": rpc error: code = NotFound desc = could not find container \"8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59\": container with ID starting with 8b4bbc8b79993b4e6f9601b8da9dc00f5d0077f13fba8fe8a7debb7b6bdb4f59 not found: ID does not exist" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729153 4888 scope.go:117] "RemoveContainer" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729316 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729334 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="74731cdc-25ca-44c5-a01b-ebe728556384" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729345 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b6f3b1-832a-4340-92dd-55b7e2797d49" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729351 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="sg-core" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729365 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729379 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="df146d80-370a-4eac-afb5-35e747ed7007" containerName="mariadb-account-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729388 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-central-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729400 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="ceilometer-notification-agent" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729408 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa347cb-768e-4e7b-ba47-42f178dec4ab" containerName="mariadb-database-create" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729417 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" containerName="proxy-httpd" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.729425 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" containerName="heat-engine" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.731381 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.737943 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.742020 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.753278 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.781211 4888 scope.go:117] "RemoveContainer" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.783578 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875\": container with ID starting with 699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875 not found: ID does not exist" containerID="699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.783633 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875"} err="failed to get container status \"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875\": rpc error: code = NotFound desc = could not find container \"699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875\": container with ID starting with 699bc04d9a243ca88524a64ae0a8550245aa08127e20c07715f6b158833f8875 not found: ID does not exist" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.808410 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.808716 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qr9\" (UniqueName: \"kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.808856 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.808988 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.809105 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.809267 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.809403 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.860604 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:52 crc kubenswrapper[4888]: E1124 00:46:52.861608 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-m5qr9 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="6c5fc3a7-83af-46e5-8434-7c8270a223ef" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911077 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911128 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qr9\" (UniqueName: \"kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911156 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911190 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911218 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911245 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911293 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.911803 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.912331 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.916626 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.916974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.917003 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.917000 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:52 crc kubenswrapper[4888]: I1124 00:46:52.930434 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qr9\" (UniqueName: \"kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9\") pod \"ceilometer-0\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " pod="openstack/ceilometer-0" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.623420 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.641432 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726135 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726275 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5qr9\" (UniqueName: \"kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726312 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726329 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726656 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726720 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.726757 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.727106 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd\") pod \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\" (UID: \"6c5fc3a7-83af-46e5-8434-7c8270a223ef\") " Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.727665 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.728015 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.731870 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.732803 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data" (OuterVolumeSpecName: "config-data") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.732837 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts" (OuterVolumeSpecName: "scripts") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.734245 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.734931 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9" (OuterVolumeSpecName: "kube-api-access-m5qr9") pod "6c5fc3a7-83af-46e5-8434-7c8270a223ef" (UID: "6c5fc3a7-83af-46e5-8434-7c8270a223ef"). InnerVolumeSpecName "kube-api-access-m5qr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829911 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829951 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829962 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829970 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc3a7-83af-46e5-8434-7c8270a223ef-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829979 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5fc3a7-83af-46e5-8434-7c8270a223ef-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:53 crc kubenswrapper[4888]: I1124 00:46:53.829987 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5qr9\" (UniqueName: \"kubernetes.io/projected/6c5fc3a7-83af-46e5-8434-7c8270a223ef-kube-api-access-m5qr9\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.255348 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241b6970-6ac3-49b0-8b51-8f727991f324" path="/var/lib/kubelet/pods/241b6970-6ac3-49b0-8b51-8f727991f324/volumes" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.256430 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acfa2cf3-151c-45de-8da2-a77724d90ee7" path="/var/lib/kubelet/pods/acfa2cf3-151c-45de-8da2-a77724d90ee7/volumes" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.634571 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.686924 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.707968 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.725206 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.728766 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.731465 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.731722 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.754918 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hw4l\" (UniqueName: \"kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853549 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853612 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853650 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853682 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853760 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.853854 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.947385 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:54 crc kubenswrapper[4888]: E1124 00:46:54.948292 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-6hw4l log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="c8d113e0-25e1-47ff-ae80-845b679d854f" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956058 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956148 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hw4l\" (UniqueName: \"kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956191 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956243 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956279 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956311 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.956406 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.957027 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.957180 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.964747 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.967374 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.969937 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.971596 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:54 crc kubenswrapper[4888]: I1124 00:46:54.981146 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hw4l\" (UniqueName: \"kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l\") pod \"ceilometer-0\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " pod="openstack/ceilometer-0" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.643653 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.657408 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.771878 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.771998 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772023 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772090 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772155 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hw4l\" (UniqueName: \"kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772203 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772250 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle\") pod \"c8d113e0-25e1-47ff-ae80-845b679d854f\" (UID: \"c8d113e0-25e1-47ff-ae80-845b679d854f\") " Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772271 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.772445 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.773042 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.773066 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8d113e0-25e1-47ff-ae80-845b679d854f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.776896 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts" (OuterVolumeSpecName: "scripts") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.778051 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l" (OuterVolumeSpecName: "kube-api-access-6hw4l") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "kube-api-access-6hw4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.778986 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.803446 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.808924 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data" (OuterVolumeSpecName: "config-data") pod "c8d113e0-25e1-47ff-ae80-845b679d854f" (UID: "c8d113e0-25e1-47ff-ae80-845b679d854f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.874484 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.874516 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.874526 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hw4l\" (UniqueName: \"kubernetes.io/projected/c8d113e0-25e1-47ff-ae80-845b679d854f-kube-api-access-6hw4l\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.874536 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:55 crc kubenswrapper[4888]: I1124 00:46:55.874544 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8d113e0-25e1-47ff-ae80-845b679d854f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.262739 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5fc3a7-83af-46e5-8434-7c8270a223ef" path="/var/lib/kubelet/pods/6c5fc3a7-83af-46e5-8434-7c8270a223ef/volumes" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.605799 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fz6k2"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.606961 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.616214 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.616272 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.616325 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tzxcg" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.619745 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fz6k2"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.651765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.691369 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.691453 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffcsk\" (UniqueName: \"kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.691495 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.691518 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.756190 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.775088 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.785014 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.788366 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.791025 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.791263 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.792580 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.793652 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.793723 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffcsk\" (UniqueName: \"kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.793764 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.793789 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.798595 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.799012 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.843522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.844893 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffcsk\" (UniqueName: \"kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk\") pod \"nova-cell0-conductor-db-sync-fz6k2\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.896136 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.896454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.896577 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.896729 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.896931 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.897089 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.897252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4czvq\" (UniqueName: \"kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.928301 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999053 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999145 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999192 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999247 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999306 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999345 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:56 crc kubenswrapper[4888]: I1124 00:46:56.999402 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4czvq\" (UniqueName: \"kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.000339 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.000673 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.004037 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.005844 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.006238 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.006293 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.021663 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4czvq\" (UniqueName: \"kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq\") pod \"ceilometer-0\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.229938 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.477912 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fz6k2"] Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.661185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" event={"ID":"141329ed-8501-47c4-bf4e-02bcde53a932","Type":"ContainerStarted","Data":"4dd5107e5b2851eac7098f508f5db90a4c7bc149fc67c7b81547d120377eba0c"} Nov 24 00:46:57 crc kubenswrapper[4888]: W1124 00:46:57.712968 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod709ec1c5_dee6_4fd0_aefd_72b5fbe380e2.slice/crio-bda0c550c1a3685b400fad15ea537c65742cd59ed863287c657d301dc9cbf0ea WatchSource:0}: Error finding container bda0c550c1a3685b400fad15ea537c65742cd59ed863287c657d301dc9cbf0ea: Status 404 returned error can't find the container with id bda0c550c1a3685b400fad15ea537c65742cd59ed863287c657d301dc9cbf0ea Nov 24 00:46:57 crc kubenswrapper[4888]: I1124 00:46:57.722225 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:46:58 crc kubenswrapper[4888]: I1124 00:46:58.259794 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d113e0-25e1-47ff-ae80-845b679d854f" path="/var/lib/kubelet/pods/c8d113e0-25e1-47ff-ae80-845b679d854f/volumes" Nov 24 00:46:58 crc kubenswrapper[4888]: I1124 00:46:58.671506 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerStarted","Data":"5123b0d5091e45f41b88795faee318e779449b00cd037518504fc649766e3dc7"} Nov 24 00:46:58 crc kubenswrapper[4888]: I1124 00:46:58.671931 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerStarted","Data":"bda0c550c1a3685b400fad15ea537c65742cd59ed863287c657d301dc9cbf0ea"} Nov 24 00:46:59 crc kubenswrapper[4888]: I1124 00:46:59.689598 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerStarted","Data":"e6d8dd6f530ef59daa3ac2eede7ae3ca0933f493c70a73f9a9a4d1123243e156"} Nov 24 00:47:00 crc kubenswrapper[4888]: I1124 00:47:00.701786 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerStarted","Data":"75c1f1334ab944e4b962a3f1fa6847a969a425e21abf18fcb691447aa63e9be3"} Nov 24 00:47:05 crc kubenswrapper[4888]: I1124 00:47:05.771736 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerStarted","Data":"dcd0ffd85856371f466a95638ec1f220ba273183fe3cbfe1942d1fbccdb36bfb"} Nov 24 00:47:05 crc kubenswrapper[4888]: I1124 00:47:05.772271 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:47:05 crc kubenswrapper[4888]: I1124 00:47:05.773628 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" event={"ID":"141329ed-8501-47c4-bf4e-02bcde53a932","Type":"ContainerStarted","Data":"f6a6deafce7428a5d422cebe9f3e9791ab247bf8374d123870a82a2198d25d56"} Nov 24 00:47:05 crc kubenswrapper[4888]: I1124 00:47:05.796896 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.371287643 podStartE2EDuration="9.79687466s" podCreationTimestamp="2025-11-24 00:46:56 +0000 UTC" firstStartedPulling="2025-11-24 00:46:57.715709778 +0000 UTC m=+1320.298393822" lastFinishedPulling="2025-11-24 00:47:05.141296785 +0000 UTC m=+1327.723980839" observedRunningTime="2025-11-24 00:47:05.793747372 +0000 UTC m=+1328.376431436" watchObservedRunningTime="2025-11-24 00:47:05.79687466 +0000 UTC m=+1328.379558714" Nov 24 00:47:05 crc kubenswrapper[4888]: I1124 00:47:05.811406 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" podStartSLOduration=2.154377216 podStartE2EDuration="9.811386315s" podCreationTimestamp="2025-11-24 00:46:56 +0000 UTC" firstStartedPulling="2025-11-24 00:46:57.484907363 +0000 UTC m=+1320.067591407" lastFinishedPulling="2025-11-24 00:47:05.141916452 +0000 UTC m=+1327.724600506" observedRunningTime="2025-11-24 00:47:05.808024391 +0000 UTC m=+1328.390708445" watchObservedRunningTime="2025-11-24 00:47:05.811386315 +0000 UTC m=+1328.394070369" Nov 24 00:47:14 crc kubenswrapper[4888]: I1124 00:47:14.888927 4888 generic.go:334] "Generic (PLEG): container finished" podID="141329ed-8501-47c4-bf4e-02bcde53a932" containerID="f6a6deafce7428a5d422cebe9f3e9791ab247bf8374d123870a82a2198d25d56" exitCode=0 Nov 24 00:47:14 crc kubenswrapper[4888]: I1124 00:47:14.888984 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" event={"ID":"141329ed-8501-47c4-bf4e-02bcde53a932","Type":"ContainerDied","Data":"f6a6deafce7428a5d422cebe9f3e9791ab247bf8374d123870a82a2198d25d56"} Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.300068 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.396422 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffcsk\" (UniqueName: \"kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk\") pod \"141329ed-8501-47c4-bf4e-02bcde53a932\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.396547 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data\") pod \"141329ed-8501-47c4-bf4e-02bcde53a932\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.396620 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle\") pod \"141329ed-8501-47c4-bf4e-02bcde53a932\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.396784 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts\") pod \"141329ed-8501-47c4-bf4e-02bcde53a932\" (UID: \"141329ed-8501-47c4-bf4e-02bcde53a932\") " Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.404211 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts" (OuterVolumeSpecName: "scripts") pod "141329ed-8501-47c4-bf4e-02bcde53a932" (UID: "141329ed-8501-47c4-bf4e-02bcde53a932"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.404252 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk" (OuterVolumeSpecName: "kube-api-access-ffcsk") pod "141329ed-8501-47c4-bf4e-02bcde53a932" (UID: "141329ed-8501-47c4-bf4e-02bcde53a932"). InnerVolumeSpecName "kube-api-access-ffcsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.431635 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "141329ed-8501-47c4-bf4e-02bcde53a932" (UID: "141329ed-8501-47c4-bf4e-02bcde53a932"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.436057 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data" (OuterVolumeSpecName: "config-data") pod "141329ed-8501-47c4-bf4e-02bcde53a932" (UID: "141329ed-8501-47c4-bf4e-02bcde53a932"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.504778 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.504861 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffcsk\" (UniqueName: \"kubernetes.io/projected/141329ed-8501-47c4-bf4e-02bcde53a932-kube-api-access-ffcsk\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.504890 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.504905 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/141329ed-8501-47c4-bf4e-02bcde53a932-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.912335 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" event={"ID":"141329ed-8501-47c4-bf4e-02bcde53a932","Type":"ContainerDied","Data":"4dd5107e5b2851eac7098f508f5db90a4c7bc149fc67c7b81547d120377eba0c"} Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.912727 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dd5107e5b2851eac7098f508f5db90a4c7bc149fc67c7b81547d120377eba0c" Nov 24 00:47:16 crc kubenswrapper[4888]: I1124 00:47:16.912527 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fz6k2" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.020769 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 00:47:17 crc kubenswrapper[4888]: E1124 00:47:17.021403 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141329ed-8501-47c4-bf4e-02bcde53a932" containerName="nova-cell0-conductor-db-sync" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.021426 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="141329ed-8501-47c4-bf4e-02bcde53a932" containerName="nova-cell0-conductor-db-sync" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.021697 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="141329ed-8501-47c4-bf4e-02bcde53a932" containerName="nova-cell0-conductor-db-sync" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.022728 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.025621 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tzxcg" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.025844 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.035712 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.116168 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.116297 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrrg\" (UniqueName: \"kubernetes.io/projected/be8f5419-cba3-4446-937d-f8d4b06c1c0e-kube-api-access-vqrrg\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.116346 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.218216 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrrg\" (UniqueName: \"kubernetes.io/projected/be8f5419-cba3-4446-937d-f8d4b06c1c0e-kube-api-access-vqrrg\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.218321 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.218394 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.223476 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.225299 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be8f5419-cba3-4446-937d-f8d4b06c1c0e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.235218 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrrg\" (UniqueName: \"kubernetes.io/projected/be8f5419-cba3-4446-937d-f8d4b06c1c0e-kube-api-access-vqrrg\") pod \"nova-cell0-conductor-0\" (UID: \"be8f5419-cba3-4446-937d-f8d4b06c1c0e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.358477 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:17 crc kubenswrapper[4888]: W1124 00:47:17.843002 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe8f5419_cba3_4446_937d_f8d4b06c1c0e.slice/crio-e5cd53bbad85e2d3aa14739eb5a6863123a5676c3a8bb46a67d6a35a96e0d030 WatchSource:0}: Error finding container e5cd53bbad85e2d3aa14739eb5a6863123a5676c3a8bb46a67d6a35a96e0d030: Status 404 returned error can't find the container with id e5cd53bbad85e2d3aa14739eb5a6863123a5676c3a8bb46a67d6a35a96e0d030 Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.843394 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 00:47:17 crc kubenswrapper[4888]: I1124 00:47:17.922648 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"be8f5419-cba3-4446-937d-f8d4b06c1c0e","Type":"ContainerStarted","Data":"e5cd53bbad85e2d3aa14739eb5a6863123a5676c3a8bb46a67d6a35a96e0d030"} Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.493054 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.493632 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-central-agent" containerID="cri-o://5123b0d5091e45f41b88795faee318e779449b00cd037518504fc649766e3dc7" gracePeriod=30 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.493716 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="sg-core" containerID="cri-o://75c1f1334ab944e4b962a3f1fa6847a969a425e21abf18fcb691447aa63e9be3" gracePeriod=30 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.493797 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-notification-agent" containerID="cri-o://e6d8dd6f530ef59daa3ac2eede7ae3ca0933f493c70a73f9a9a4d1123243e156" gracePeriod=30 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.493787 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="proxy-httpd" containerID="cri-o://dcd0ffd85856371f466a95638ec1f220ba273183fe3cbfe1942d1fbccdb36bfb" gracePeriod=30 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.505288 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.215:3000/\": EOF" Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.933133 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"be8f5419-cba3-4446-937d-f8d4b06c1c0e","Type":"ContainerStarted","Data":"50610f8a073ae57c861a2047665f24044564c3860285f64d8741820bd97d268c"} Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.933261 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.935939 4888 generic.go:334] "Generic (PLEG): container finished" podID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerID="dcd0ffd85856371f466a95638ec1f220ba273183fe3cbfe1942d1fbccdb36bfb" exitCode=0 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.935965 4888 generic.go:334] "Generic (PLEG): container finished" podID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerID="75c1f1334ab944e4b962a3f1fa6847a969a425e21abf18fcb691447aa63e9be3" exitCode=2 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.935972 4888 generic.go:334] "Generic (PLEG): container finished" podID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerID="5123b0d5091e45f41b88795faee318e779449b00cd037518504fc649766e3dc7" exitCode=0 Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.935994 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerDied","Data":"dcd0ffd85856371f466a95638ec1f220ba273183fe3cbfe1942d1fbccdb36bfb"} Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.936018 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerDied","Data":"75c1f1334ab944e4b962a3f1fa6847a969a425e21abf18fcb691447aa63e9be3"} Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.936030 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerDied","Data":"5123b0d5091e45f41b88795faee318e779449b00cd037518504fc649766e3dc7"} Nov 24 00:47:18 crc kubenswrapper[4888]: I1124 00:47:18.949761 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.949743217 podStartE2EDuration="1.949743217s" podCreationTimestamp="2025-11-24 00:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:18.945932831 +0000 UTC m=+1341.528616875" watchObservedRunningTime="2025-11-24 00:47:18.949743217 +0000 UTC m=+1341.532427261" Nov 24 00:47:19 crc kubenswrapper[4888]: I1124 00:47:19.973451 4888 generic.go:334] "Generic (PLEG): container finished" podID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerID="e6d8dd6f530ef59daa3ac2eede7ae3ca0933f493c70a73f9a9a4d1123243e156" exitCode=0 Nov 24 00:47:19 crc kubenswrapper[4888]: I1124 00:47:19.973537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerDied","Data":"e6d8dd6f530ef59daa3ac2eede7ae3ca0933f493c70a73f9a9a4d1123243e156"} Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.309425 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.486620 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.486838 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.486873 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.486925 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.486966 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.487049 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4czvq\" (UniqueName: \"kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.487108 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml\") pod \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\" (UID: \"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2\") " Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.487388 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.487568 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.487715 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.493100 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq" (OuterVolumeSpecName: "kube-api-access-4czvq") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "kube-api-access-4czvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.494009 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts" (OuterVolumeSpecName: "scripts") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.519026 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.589311 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.589344 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.589358 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4czvq\" (UniqueName: \"kubernetes.io/projected/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-kube-api-access-4czvq\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.589371 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.604669 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.620743 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data" (OuterVolumeSpecName: "config-data") pod "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" (UID: "709ec1c5-dee6-4fd0-aefd-72b5fbe380e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.691731 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.691774 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.983793 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"709ec1c5-dee6-4fd0-aefd-72b5fbe380e2","Type":"ContainerDied","Data":"bda0c550c1a3685b400fad15ea537c65742cd59ed863287c657d301dc9cbf0ea"} Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.984044 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:20 crc kubenswrapper[4888]: I1124 00:47:20.984141 4888 scope.go:117] "RemoveContainer" containerID="dcd0ffd85856371f466a95638ec1f220ba273183fe3cbfe1942d1fbccdb36bfb" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.005834 4888 scope.go:117] "RemoveContainer" containerID="75c1f1334ab944e4b962a3f1fa6847a969a425e21abf18fcb691447aa63e9be3" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.026985 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.029074 4888 scope.go:117] "RemoveContainer" containerID="e6d8dd6f530ef59daa3ac2eede7ae3ca0933f493c70a73f9a9a4d1123243e156" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.051658 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.060398 4888 scope.go:117] "RemoveContainer" containerID="5123b0d5091e45f41b88795faee318e779449b00cd037518504fc649766e3dc7" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.068622 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:21 crc kubenswrapper[4888]: E1124 00:47:21.069133 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="sg-core" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069147 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="sg-core" Nov 24 00:47:21 crc kubenswrapper[4888]: E1124 00:47:21.069162 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="proxy-httpd" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069169 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="proxy-httpd" Nov 24 00:47:21 crc kubenswrapper[4888]: E1124 00:47:21.069190 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-central-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069198 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-central-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: E1124 00:47:21.069223 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-notification-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069228 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-notification-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069421 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-notification-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069440 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="proxy-httpd" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069450 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="sg-core" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.069468 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" containerName="ceilometer-central-agent" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.071595 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.074317 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.074681 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.086122 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200050 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200115 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j995z\" (UniqueName: \"kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200269 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200478 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200685 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200768 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.200915 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303358 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303485 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j995z\" (UniqueName: \"kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303702 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303777 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303860 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303915 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.303865 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.304001 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.304228 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.309981 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.310712 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.317566 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.318471 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.328856 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j995z\" (UniqueName: \"kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z\") pod \"ceilometer-0\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.411105 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:21 crc kubenswrapper[4888]: I1124 00:47:21.955249 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:22 crc kubenswrapper[4888]: I1124 00:47:22.003543 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerStarted","Data":"a3f441832521a95294c215a6fadf36135949d0d5c19a5001fd3e996ea081117e"} Nov 24 00:47:22 crc kubenswrapper[4888]: I1124 00:47:22.261831 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709ec1c5-dee6-4fd0-aefd-72b5fbe380e2" path="/var/lib/kubelet/pods/709ec1c5-dee6-4fd0-aefd-72b5fbe380e2/volumes" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.014025 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerStarted","Data":"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31"} Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.134468 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-a098-account-create-7tggp"] Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.136429 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.139486 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.151503 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-fx726"] Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.152971 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.165006 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-a098-account-create-7tggp"] Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.187904 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-fx726"] Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.264003 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4sh\" (UniqueName: \"kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.264197 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.264252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn6tn\" (UniqueName: \"kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.264284 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.369007 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn6tn\" (UniqueName: \"kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.369259 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.369364 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4sh\" (UniqueName: \"kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.369606 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.370236 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.370434 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.403523 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4sh\" (UniqueName: \"kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh\") pod \"aodh-db-create-fx726\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " pod="openstack/aodh-db-create-fx726" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.415614 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn6tn\" (UniqueName: \"kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn\") pod \"aodh-a098-account-create-7tggp\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.602092 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:23 crc kubenswrapper[4888]: I1124 00:47:23.611354 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-fx726" Nov 24 00:47:24 crc kubenswrapper[4888]: I1124 00:47:24.023867 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerStarted","Data":"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa"} Nov 24 00:47:24 crc kubenswrapper[4888]: I1124 00:47:24.193248 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-a098-account-create-7tggp"] Nov 24 00:47:24 crc kubenswrapper[4888]: I1124 00:47:24.307725 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-fx726"] Nov 24 00:47:24 crc kubenswrapper[4888]: W1124 00:47:24.321474 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb23b2a93_fe3a_4fff_a78e_13a50826aefb.slice/crio-3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727 WatchSource:0}: Error finding container 3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727: Status 404 returned error can't find the container with id 3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727 Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.033996 4888 generic.go:334] "Generic (PLEG): container finished" podID="43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" containerID="0a19c69bb1abb58b20426b7b81a3e6f3de3a6aa5a8d50a4bbbe40fdabf8bfafd" exitCode=0 Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.034051 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a098-account-create-7tggp" event={"ID":"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed","Type":"ContainerDied","Data":"0a19c69bb1abb58b20426b7b81a3e6f3de3a6aa5a8d50a4bbbe40fdabf8bfafd"} Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.034386 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a098-account-create-7tggp" event={"ID":"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed","Type":"ContainerStarted","Data":"11a6d1f6f1bc9be4241de4433fa04b0c4bfedf1434c5451bdc7ed6c3def41580"} Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.035932 4888 generic.go:334] "Generic (PLEG): container finished" podID="b23b2a93-fe3a-4fff-a78e-13a50826aefb" containerID="e896c578b132a12412e9a49f777d8686b34a6d316ffe3e72d39f23387f5806c7" exitCode=0 Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.036004 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-fx726" event={"ID":"b23b2a93-fe3a-4fff-a78e-13a50826aefb","Type":"ContainerDied","Data":"e896c578b132a12412e9a49f777d8686b34a6d316ffe3e72d39f23387f5806c7"} Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.036030 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-fx726" event={"ID":"b23b2a93-fe3a-4fff-a78e-13a50826aefb","Type":"ContainerStarted","Data":"3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727"} Nov 24 00:47:25 crc kubenswrapper[4888]: I1124 00:47:25.037996 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerStarted","Data":"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835"} Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.051019 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerStarted","Data":"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1"} Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.086614 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.472779393 podStartE2EDuration="5.086590729s" podCreationTimestamp="2025-11-24 00:47:21 +0000 UTC" firstStartedPulling="2025-11-24 00:47:21.964400936 +0000 UTC m=+1344.547084980" lastFinishedPulling="2025-11-24 00:47:25.578212252 +0000 UTC m=+1348.160896316" observedRunningTime="2025-11-24 00:47:26.085246282 +0000 UTC m=+1348.667930326" watchObservedRunningTime="2025-11-24 00:47:26.086590729 +0000 UTC m=+1348.669274783" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.494203 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.645115 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-fx726" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.649326 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn6tn\" (UniqueName: \"kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn\") pod \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.650254 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts\") pod \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\" (UID: \"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed\") " Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.650990 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" (UID: "43af5b1c-1c23-41ea-b90a-f2025b4bb8ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.656030 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn" (OuterVolumeSpecName: "kube-api-access-pn6tn") pod "43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" (UID: "43af5b1c-1c23-41ea-b90a-f2025b4bb8ed"). InnerVolumeSpecName "kube-api-access-pn6tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.752323 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l4sh\" (UniqueName: \"kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh\") pod \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.752601 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts\") pod \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\" (UID: \"b23b2a93-fe3a-4fff-a78e-13a50826aefb\") " Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.753203 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn6tn\" (UniqueName: \"kubernetes.io/projected/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-kube-api-access-pn6tn\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.753222 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.753269 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b23b2a93-fe3a-4fff-a78e-13a50826aefb" (UID: "b23b2a93-fe3a-4fff-a78e-13a50826aefb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.756037 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh" (OuterVolumeSpecName: "kube-api-access-6l4sh") pod "b23b2a93-fe3a-4fff-a78e-13a50826aefb" (UID: "b23b2a93-fe3a-4fff-a78e-13a50826aefb"). InnerVolumeSpecName "kube-api-access-6l4sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.854869 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l4sh\" (UniqueName: \"kubernetes.io/projected/b23b2a93-fe3a-4fff-a78e-13a50826aefb-kube-api-access-6l4sh\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:26 crc kubenswrapper[4888]: I1124 00:47:26.854900 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b23b2a93-fe3a-4fff-a78e-13a50826aefb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.061747 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-a098-account-create-7tggp" event={"ID":"43af5b1c-1c23-41ea-b90a-f2025b4bb8ed","Type":"ContainerDied","Data":"11a6d1f6f1bc9be4241de4433fa04b0c4bfedf1434c5451bdc7ed6c3def41580"} Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.061794 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11a6d1f6f1bc9be4241de4433fa04b0c4bfedf1434c5451bdc7ed6c3def41580" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.061768 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-a098-account-create-7tggp" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.079232 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-fx726" event={"ID":"b23b2a93-fe3a-4fff-a78e-13a50826aefb","Type":"ContainerDied","Data":"3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727"} Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.079268 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-fx726" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.079273 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3815ca99ae1a29f96adc79f3e8f266f4accff030d20b8d54d59ff3f3ce97e727" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.079377 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.410353 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.960312 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vd92r"] Nov 24 00:47:27 crc kubenswrapper[4888]: E1124 00:47:27.961463 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b23b2a93-fe3a-4fff-a78e-13a50826aefb" containerName="mariadb-database-create" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.961492 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b23b2a93-fe3a-4fff-a78e-13a50826aefb" containerName="mariadb-database-create" Nov 24 00:47:27 crc kubenswrapper[4888]: E1124 00:47:27.961517 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" containerName="mariadb-account-create" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.961526 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" containerName="mariadb-account-create" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.961767 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" containerName="mariadb-account-create" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.961798 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b23b2a93-fe3a-4fff-a78e-13a50826aefb" containerName="mariadb-database-create" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.962726 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.967227 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.968203 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 00:47:27 crc kubenswrapper[4888]: I1124 00:47:27.980423 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vd92r"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.111601 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.111728 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.111781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xc5b\" (UniqueName: \"kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.111823 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.124555 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.125933 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.131508 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.141720 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.215098 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.215189 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xc5b\" (UniqueName: \"kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.215218 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.215281 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.227594 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.244599 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.249419 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.280466 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xc5b\" (UniqueName: \"kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b\") pod \"nova-cell0-cell-mapping-vd92r\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.288757 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.322187 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flz28\" (UniqueName: \"kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.323878 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.324294 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.341086 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.342574 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.343771 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.343968 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.358551 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.370295 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.370522 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.394873 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425106 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425546 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425631 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425663 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425696 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flz28\" (UniqueName: \"kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425727 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425745 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425822 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425858 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgqhf\" (UniqueName: \"kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425888 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88s2z\" (UniqueName: \"kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.425910 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.429265 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.431196 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.432218 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.439891 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.441893 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.447719 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.449439 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.455635 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.466804 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flz28\" (UniqueName: \"kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28\") pod \"nova-scheduler-0\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527207 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527265 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgqhf\" (UniqueName: \"kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527299 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88s2z\" (UniqueName: \"kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527343 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527375 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527400 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.527443 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.529863 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.537838 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.541420 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.545462 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.563039 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.572496 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88s2z\" (UniqueName: \"kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.584158 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgqhf\" (UniqueName: \"kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf\") pod \"nova-metadata-0\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630063 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630343 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630363 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98jzx\" (UniqueName: \"kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630381 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630427 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630447 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630476 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f65nm\" (UniqueName: \"kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630502 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630523 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.630559 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.688975 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.718131 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738106 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738197 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738232 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738249 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98jzx\" (UniqueName: \"kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738265 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738307 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738324 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738360 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f65nm\" (UniqueName: \"kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738388 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.738417 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.750079 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.751044 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-w7s5d"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.751973 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.752331 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.753597 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.771545 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.771750 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.771888 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.772726 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.773051 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-t9vsv" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.784207 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.784542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.785174 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.803364 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.809502 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-w7s5d"] Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.810472 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98jzx\" (UniqueName: \"kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.816191 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f65nm\" (UniqueName: \"kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm\") pod \"dnsmasq-dns-9b86998b5-cwd26\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.828675 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.833579 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data\") pod \"nova-api-0\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " pod="openstack/nova-api-0" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.946749 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.946827 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.946863 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wfgq\" (UniqueName: \"kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:28 crc kubenswrapper[4888]: I1124 00:47:28.946927 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.058221 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.058593 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.058630 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.058653 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wfgq\" (UniqueName: \"kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.063558 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.067612 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.079848 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.081463 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.098031 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wfgq\" (UniqueName: \"kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq\") pod \"aodh-db-sync-w7s5d\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.166804 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vd92r"] Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.267600 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:29 crc kubenswrapper[4888]: W1124 00:47:29.300896 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a3e9761_6aaf_43dd_ac54_d67a186b47b6.slice/crio-742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d WatchSource:0}: Error finding container 742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d: Status 404 returned error can't find the container with id 742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.908909 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.947363 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgpfg"] Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.948727 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.957283 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.962490 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 00:47:29 crc kubenswrapper[4888]: I1124 00:47:29.989525 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgpfg"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.014519 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.043648 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.092553 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lpz6\" (UniqueName: \"kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.092587 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.095137 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.095293 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.115425 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.127517 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-w7s5d"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.140691 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.176797 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe","Type":"ContainerStarted","Data":"366cf529b5c956c6e7ac1e3b32b0408c146f93deffb6e0237a22c851ab0ff8ff"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.181728 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerStarted","Data":"a4094de7a7432a294be324c4d15ce5eeac2f275522fa4a0e449517feb71596d8"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.183157 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39c5c954-3e9f-425f-b250-bf1bdbbb7d84","Type":"ContainerStarted","Data":"2adf0eef4e0dd78a23419487f9c3541c4fada338a5c8c83e7adf0130ca82b346"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.183956 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" event={"ID":"53a42b6c-9404-41d4-9d05-5c30e11fece7","Type":"ContainerStarted","Data":"99fc1bf87533e93afabd5302abbf3fe0497abcd4ce08b107b3624495b01929f5"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.184707 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w7s5d" event={"ID":"166b8865-2142-4a68-852b-ac7511643e32","Type":"ContainerStarted","Data":"7498972bf23c00601c9dcff631465dc1e7db699280ae1ea251a8df7e7e1af9fd"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.186107 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vd92r" event={"ID":"1a3e9761-6aaf-43dd-ac54-d67a186b47b6","Type":"ContainerStarted","Data":"6f7dd1280ac78f8578d663849476be4e1ae40e0cabb40af6ab076efaa519f3d8"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.186129 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vd92r" event={"ID":"1a3e9761-6aaf-43dd-ac54-d67a186b47b6","Type":"ContainerStarted","Data":"742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d"} Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.198247 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.198339 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lpz6\" (UniqueName: \"kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.198366 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.198448 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.208638 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.208907 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.218440 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.229590 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vd92r" podStartSLOduration=3.229569474 podStartE2EDuration="3.229569474s" podCreationTimestamp="2025-11-24 00:47:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:30.202849557 +0000 UTC m=+1352.785533601" watchObservedRunningTime="2025-11-24 00:47:30.229569474 +0000 UTC m=+1352.812253518" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.231094 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lpz6\" (UniqueName: \"kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6\") pod \"nova-cell1-conductor-db-sync-vgpfg\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.389388 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:30 crc kubenswrapper[4888]: I1124 00:47:30.870996 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgpfg"] Nov 24 00:47:30 crc kubenswrapper[4888]: W1124 00:47:30.895766 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02a71bf4_35a2_4442_b9e5_828e553ad5e1.slice/crio-e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8 WatchSource:0}: Error finding container e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8: Status 404 returned error can't find the container with id e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8 Nov 24 00:47:31 crc kubenswrapper[4888]: I1124 00:47:31.203797 4888 generic.go:334] "Generic (PLEG): container finished" podID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerID="baa1f270019b3951eff8163acdaeba4ba62e88a91db8323106434568c5ffa895" exitCode=0 Nov 24 00:47:31 crc kubenswrapper[4888]: I1124 00:47:31.204185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" event={"ID":"53a42b6c-9404-41d4-9d05-5c30e11fece7","Type":"ContainerDied","Data":"baa1f270019b3951eff8163acdaeba4ba62e88a91db8323106434568c5ffa895"} Nov 24 00:47:31 crc kubenswrapper[4888]: I1124 00:47:31.218073 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" event={"ID":"02a71bf4-35a2-4442-b9e5-828e553ad5e1","Type":"ContainerStarted","Data":"e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8"} Nov 24 00:47:31 crc kubenswrapper[4888]: I1124 00:47:31.260218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerStarted","Data":"3059de5e0f8fb591e3bc8f3fda34a0fb12d9888ff60a7428d01a75be49189c4f"} Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.270874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" event={"ID":"53a42b6c-9404-41d4-9d05-5c30e11fece7","Type":"ContainerStarted","Data":"01082314d83ea4e6e600094b33ddf8e22f486f333fba141c8b3cb9c8676bc453"} Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.271259 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.274155 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" event={"ID":"02a71bf4-35a2-4442-b9e5-828e553ad5e1","Type":"ContainerStarted","Data":"9b2bae8aaffde05723f00f49fae9acf262fd364914738f71f574e40a08039c11"} Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.289217 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" podStartSLOduration=4.289202744 podStartE2EDuration="4.289202744s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:32.287498497 +0000 UTC m=+1354.870182541" watchObservedRunningTime="2025-11-24 00:47:32.289202744 +0000 UTC m=+1354.871886788" Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.311520 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" podStartSLOduration=3.311503368 podStartE2EDuration="3.311503368s" podCreationTimestamp="2025-11-24 00:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:32.304590805 +0000 UTC m=+1354.887274849" watchObservedRunningTime="2025-11-24 00:47:32.311503368 +0000 UTC m=+1354.894187412" Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.454932 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:47:32 crc kubenswrapper[4888]: I1124 00:47:32.503916 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.355667 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerStarted","Data":"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.356222 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerStarted","Data":"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.359652 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerStarted","Data":"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.359713 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerStarted","Data":"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.359769 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-log" containerID="cri-o://e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3" gracePeriod=30 Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.359907 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-metadata" containerID="cri-o://efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24" gracePeriod=30 Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.363548 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39c5c954-3e9f-425f-b250-bf1bdbbb7d84","Type":"ContainerStarted","Data":"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.363652 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a" gracePeriod=30 Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.372881 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w7s5d" event={"ID":"166b8865-2142-4a68-852b-ac7511643e32","Type":"ContainerStarted","Data":"4e1ec1f25ca75c582b1065a410999ea908110682881e906b7332c6d976e42589"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.381424 4888 generic.go:334] "Generic (PLEG): container finished" podID="1a3e9761-6aaf-43dd-ac54-d67a186b47b6" containerID="6f7dd1280ac78f8578d663849476be4e1ae40e0cabb40af6ab076efaa519f3d8" exitCode=0 Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.381520 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vd92r" event={"ID":"1a3e9761-6aaf-43dd-ac54-d67a186b47b6","Type":"ContainerDied","Data":"6f7dd1280ac78f8578d663849476be4e1ae40e0cabb40af6ab076efaa519f3d8"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.383861 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe","Type":"ContainerStarted","Data":"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442"} Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.384922 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.120694856 podStartE2EDuration="10.38490662s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="2025-11-24 00:47:30.160022559 +0000 UTC m=+1352.742706603" lastFinishedPulling="2025-11-24 00:47:37.424234323 +0000 UTC m=+1360.006918367" observedRunningTime="2025-11-24 00:47:38.37596184 +0000 UTC m=+1360.958645904" watchObservedRunningTime="2025-11-24 00:47:38.38490662 +0000 UTC m=+1360.967590664" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.406127 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.045359459 podStartE2EDuration="10.406110743s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="2025-11-24 00:47:30.050029523 +0000 UTC m=+1352.632713567" lastFinishedPulling="2025-11-24 00:47:37.410780797 +0000 UTC m=+1359.993464851" observedRunningTime="2025-11-24 00:47:38.388706856 +0000 UTC m=+1360.971390900" watchObservedRunningTime="2025-11-24 00:47:38.406110743 +0000 UTC m=+1360.988794787" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.415226 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.954487188 podStartE2EDuration="10.415206288s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="2025-11-24 00:47:30.004011246 +0000 UTC m=+1352.586695280" lastFinishedPulling="2025-11-24 00:47:37.464730336 +0000 UTC m=+1360.047414380" observedRunningTime="2025-11-24 00:47:38.408836379 +0000 UTC m=+1360.991520423" watchObservedRunningTime="2025-11-24 00:47:38.415206288 +0000 UTC m=+1360.997890332" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.444085 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-w7s5d" podStartSLOduration=3.072732654 podStartE2EDuration="10.444053854s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="2025-11-24 00:47:30.117965733 +0000 UTC m=+1352.700649777" lastFinishedPulling="2025-11-24 00:47:37.489286933 +0000 UTC m=+1360.071970977" observedRunningTime="2025-11-24 00:47:38.434207339 +0000 UTC m=+1361.016891383" watchObservedRunningTime="2025-11-24 00:47:38.444053854 +0000 UTC m=+1361.026737888" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.524976 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.016034338 podStartE2EDuration="10.524954016s" podCreationTimestamp="2025-11-24 00:47:28 +0000 UTC" firstStartedPulling="2025-11-24 00:47:29.915095399 +0000 UTC m=+1352.497779443" lastFinishedPulling="2025-11-24 00:47:37.424015077 +0000 UTC m=+1360.006699121" observedRunningTime="2025-11-24 00:47:38.485792672 +0000 UTC m=+1361.068476716" watchObservedRunningTime="2025-11-24 00:47:38.524954016 +0000 UTC m=+1361.107638060" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.690144 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.690192 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.719335 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.754463 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.754504 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.785096 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.830109 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.900983 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:47:38 crc kubenswrapper[4888]: I1124 00:47:38.901248 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="dnsmasq-dns" containerID="cri-o://e9c8478655a21127b52ab79ea30d43fc2ca3a9c4f02745a8f438b709b73f0f3a" gracePeriod=10 Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.082263 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.082301 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.410836 4888 generic.go:334] "Generic (PLEG): container finished" podID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerID="e9c8478655a21127b52ab79ea30d43fc2ca3a9c4f02745a8f438b709b73f0f3a" exitCode=0 Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.411131 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" event={"ID":"c9b000bb-1961-4ece-b8df-13dd16d4e2fa","Type":"ContainerDied","Data":"e9c8478655a21127b52ab79ea30d43fc2ca3a9c4f02745a8f438b709b73f0f3a"} Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.424571 4888 generic.go:334] "Generic (PLEG): container finished" podID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerID="e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3" exitCode=143 Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.424870 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerDied","Data":"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3"} Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.502471 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.671653 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.772392 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.772587 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.772646 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nrk7\" (UniqueName: \"kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.772677 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.772959 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.773013 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb\") pod \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\" (UID: \"c9b000bb-1961-4ece-b8df-13dd16d4e2fa\") " Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.803092 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7" (OuterVolumeSpecName: "kube-api-access-4nrk7") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "kube-api-access-4nrk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.850957 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.859402 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.882563 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.882594 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nrk7\" (UniqueName: \"kubernetes.io/projected/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-kube-api-access-4nrk7\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.882607 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.891455 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config" (OuterVolumeSpecName: "config") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.931274 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.945320 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c9b000bb-1961-4ece-b8df-13dd16d4e2fa" (UID: "c9b000bb-1961-4ece-b8df-13dd16d4e2fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.984332 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.984371 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.984382 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9b000bb-1961-4ece-b8df-13dd16d4e2fa-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:39 crc kubenswrapper[4888]: I1124 00:47:39.986589 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.090595 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle\") pod \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.090879 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts\") pod \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.090915 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data\") pod \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.090962 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xc5b\" (UniqueName: \"kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b\") pod \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\" (UID: \"1a3e9761-6aaf-43dd-ac54-d67a186b47b6\") " Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.097935 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts" (OuterVolumeSpecName: "scripts") pod "1a3e9761-6aaf-43dd-ac54-d67a186b47b6" (UID: "1a3e9761-6aaf-43dd-ac54-d67a186b47b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.097970 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b" (OuterVolumeSpecName: "kube-api-access-4xc5b") pod "1a3e9761-6aaf-43dd-ac54-d67a186b47b6" (UID: "1a3e9761-6aaf-43dd-ac54-d67a186b47b6"). InnerVolumeSpecName "kube-api-access-4xc5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.119802 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data" (OuterVolumeSpecName: "config-data") pod "1a3e9761-6aaf-43dd-ac54-d67a186b47b6" (UID: "1a3e9761-6aaf-43dd-ac54-d67a186b47b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.123063 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.224:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.124979 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a3e9761-6aaf-43dd-ac54-d67a186b47b6" (UID: "1a3e9761-6aaf-43dd-ac54-d67a186b47b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.164082 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.224:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.193888 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.193923 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.193935 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xc5b\" (UniqueName: \"kubernetes.io/projected/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-kube-api-access-4xc5b\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.193949 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3e9761-6aaf-43dd-ac54-d67a186b47b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.435071 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vd92r" event={"ID":"1a3e9761-6aaf-43dd-ac54-d67a186b47b6","Type":"ContainerDied","Data":"742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d"} Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.435117 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="742d88788f6e0b94868b20ca9490f63ca18c06a6ec4aedf4f570ff642f25a28d" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.435183 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vd92r" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.440328 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.440937 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-brzj4" event={"ID":"c9b000bb-1961-4ece-b8df-13dd16d4e2fa","Type":"ContainerDied","Data":"1c9643c47170e49bf0df0067e3ac4755a5b1b69b0f4e554752da5b631b1993b4"} Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.440975 4888 scope.go:117] "RemoveContainer" containerID="e9c8478655a21127b52ab79ea30d43fc2ca3a9c4f02745a8f438b709b73f0f3a" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.468900 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.472902 4888 scope.go:117] "RemoveContainer" containerID="e30aa68d1d49594dbedebf3412425956f6a8c49ad68565b4b75d66881f24a1e1" Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.481563 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-brzj4"] Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.638763 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:40 crc kubenswrapper[4888]: I1124 00:47:40.655995 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.451585 4888 generic.go:334] "Generic (PLEG): container finished" podID="166b8865-2142-4a68-852b-ac7511643e32" containerID="4e1ec1f25ca75c582b1065a410999ea908110682881e906b7332c6d976e42589" exitCode=0 Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.451721 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w7s5d" event={"ID":"166b8865-2142-4a68-852b-ac7511643e32","Type":"ContainerDied","Data":"4e1ec1f25ca75c582b1065a410999ea908110682881e906b7332c6d976e42589"} Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.454304 4888 generic.go:334] "Generic (PLEG): container finished" podID="02a71bf4-35a2-4442-b9e5-828e553ad5e1" containerID="9b2bae8aaffde05723f00f49fae9acf262fd364914738f71f574e40a08039c11" exitCode=0 Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.454384 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" event={"ID":"02a71bf4-35a2-4442-b9e5-828e553ad5e1","Type":"ContainerDied","Data":"9b2bae8aaffde05723f00f49fae9acf262fd364914738f71f574e40a08039c11"} Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.455593 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-log" containerID="cri-o://f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909" gracePeriod=30 Nov 24 00:47:41 crc kubenswrapper[4888]: I1124 00:47:41.455760 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-api" containerID="cri-o://e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f" gracePeriod=30 Nov 24 00:47:42 crc kubenswrapper[4888]: I1124 00:47:42.259189 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" path="/var/lib/kubelet/pods/c9b000bb-1961-4ece-b8df-13dd16d4e2fa/volumes" Nov 24 00:47:42 crc kubenswrapper[4888]: I1124 00:47:42.474032 4888 generic.go:334] "Generic (PLEG): container finished" podID="0c013531-f714-4704-bd2b-53136d4b0db9" containerID="f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909" exitCode=143 Nov 24 00:47:42 crc kubenswrapper[4888]: I1124 00:47:42.474237 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerDied","Data":"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909"} Nov 24 00:47:42 crc kubenswrapper[4888]: I1124 00:47:42.474470 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" containerName="nova-scheduler-scheduler" containerID="cri-o://da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442" gracePeriod=30 Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.137452 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.142025 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255343 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lpz6\" (UniqueName: \"kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6\") pod \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255417 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wfgq\" (UniqueName: \"kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq\") pod \"166b8865-2142-4a68-852b-ac7511643e32\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255498 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle\") pod \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255528 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts\") pod \"166b8865-2142-4a68-852b-ac7511643e32\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255553 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts\") pod \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255574 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data\") pod \"166b8865-2142-4a68-852b-ac7511643e32\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle\") pod \"166b8865-2142-4a68-852b-ac7511643e32\" (UID: \"166b8865-2142-4a68-852b-ac7511643e32\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.255660 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data\") pod \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\" (UID: \"02a71bf4-35a2-4442-b9e5-828e553ad5e1\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.267060 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6" (OuterVolumeSpecName: "kube-api-access-9lpz6") pod "02a71bf4-35a2-4442-b9e5-828e553ad5e1" (UID: "02a71bf4-35a2-4442-b9e5-828e553ad5e1"). InnerVolumeSpecName "kube-api-access-9lpz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.268010 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts" (OuterVolumeSpecName: "scripts") pod "166b8865-2142-4a68-852b-ac7511643e32" (UID: "166b8865-2142-4a68-852b-ac7511643e32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.271607 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq" (OuterVolumeSpecName: "kube-api-access-2wfgq") pod "166b8865-2142-4a68-852b-ac7511643e32" (UID: "166b8865-2142-4a68-852b-ac7511643e32"). InnerVolumeSpecName "kube-api-access-2wfgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.273417 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts" (OuterVolumeSpecName: "scripts") pod "02a71bf4-35a2-4442-b9e5-828e553ad5e1" (UID: "02a71bf4-35a2-4442-b9e5-828e553ad5e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.300165 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data" (OuterVolumeSpecName: "config-data") pod "02a71bf4-35a2-4442-b9e5-828e553ad5e1" (UID: "02a71bf4-35a2-4442-b9e5-828e553ad5e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.309913 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "166b8865-2142-4a68-852b-ac7511643e32" (UID: "166b8865-2142-4a68-852b-ac7511643e32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.311970 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data" (OuterVolumeSpecName: "config-data") pod "166b8865-2142-4a68-852b-ac7511643e32" (UID: "166b8865-2142-4a68-852b-ac7511643e32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.328644 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02a71bf4-35a2-4442-b9e5-828e553ad5e1" (UID: "02a71bf4-35a2-4442-b9e5-828e553ad5e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.351675 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360149 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360173 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360183 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360191 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360203 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/166b8865-2142-4a68-852b-ac7511643e32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360213 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a71bf4-35a2-4442-b9e5-828e553ad5e1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360222 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lpz6\" (UniqueName: \"kubernetes.io/projected/02a71bf4-35a2-4442-b9e5-828e553ad5e1-kube-api-access-9lpz6\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.360230 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wfgq\" (UniqueName: \"kubernetes.io/projected/166b8865-2142-4a68-852b-ac7511643e32-kube-api-access-2wfgq\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.461823 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data\") pod \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.461987 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle\") pod \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.462032 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flz28\" (UniqueName: \"kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28\") pod \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\" (UID: \"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe\") " Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.465346 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28" (OuterVolumeSpecName: "kube-api-access-flz28") pod "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" (UID: "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe"). InnerVolumeSpecName "kube-api-access-flz28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.487585 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" (UID: "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.487668 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data" (OuterVolumeSpecName: "config-data") pod "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" (UID: "4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.501074 4888 generic.go:334] "Generic (PLEG): container finished" podID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" containerID="da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442" exitCode=0 Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.501148 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe","Type":"ContainerDied","Data":"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442"} Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.501174 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe","Type":"ContainerDied","Data":"366cf529b5c956c6e7ac1e3b32b0408c146f93deffb6e0237a22c851ab0ff8ff"} Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.501189 4888 scope.go:117] "RemoveContainer" containerID="da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.502353 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.505711 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w7s5d" event={"ID":"166b8865-2142-4a68-852b-ac7511643e32","Type":"ContainerDied","Data":"7498972bf23c00601c9dcff631465dc1e7db699280ae1ea251a8df7e7e1af9fd"} Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.505750 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7498972bf23c00601c9dcff631465dc1e7db699280ae1ea251a8df7e7e1af9fd" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.505939 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w7s5d" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.511508 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" event={"ID":"02a71bf4-35a2-4442-b9e5-828e553ad5e1","Type":"ContainerDied","Data":"e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8"} Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.511640 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e87ad5913dd7aeb1008c3510c53449524e29be4b200a49f20ee5bcfc1e0cdcb8" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.511791 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgpfg" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.529682 4888 scope.go:117] "RemoveContainer" containerID="da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.530543 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442\": container with ID starting with da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442 not found: ID does not exist" containerID="da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.530584 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442"} err="failed to get container status \"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442\": rpc error: code = NotFound desc = could not find container \"da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442\": container with ID starting with da668d383b8364988cfacf9ecc6a25000d2e7eb4c0fd9214095e95fde20c6442 not found: ID does not exist" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.564244 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.564498 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.564572 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flz28\" (UniqueName: \"kubernetes.io/projected/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe-kube-api-access-flz28\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.567030 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.580331 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.595997 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596388 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166b8865-2142-4a68-852b-ac7511643e32" containerName="aodh-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596404 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="166b8865-2142-4a68-852b-ac7511643e32" containerName="aodh-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596416 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" containerName="nova-scheduler-scheduler" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596427 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" containerName="nova-scheduler-scheduler" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596443 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a3e9761-6aaf-43dd-ac54-d67a186b47b6" containerName="nova-manage" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596450 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a3e9761-6aaf-43dd-ac54-d67a186b47b6" containerName="nova-manage" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596459 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a71bf4-35a2-4442-b9e5-828e553ad5e1" containerName="nova-cell1-conductor-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596465 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a71bf4-35a2-4442-b9e5-828e553ad5e1" containerName="nova-cell1-conductor-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596488 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="dnsmasq-dns" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596494 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="dnsmasq-dns" Nov 24 00:47:43 crc kubenswrapper[4888]: E1124 00:47:43.596516 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="init" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596521 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="init" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596704 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" containerName="nova-scheduler-scheduler" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596718 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b000bb-1961-4ece-b8df-13dd16d4e2fa" containerName="dnsmasq-dns" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596729 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a3e9761-6aaf-43dd-ac54-d67a186b47b6" containerName="nova-manage" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596738 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a71bf4-35a2-4442-b9e5-828e553ad5e1" containerName="nova-cell1-conductor-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.596752 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="166b8865-2142-4a68-852b-ac7511643e32" containerName="aodh-db-sync" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.597369 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.597442 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.614692 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.676411 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.685798 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.692751 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.708314 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.767751 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8cwv\" (UniqueName: \"kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.767796 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.768207 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.800682 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.806224 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.809299 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.809652 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-t9vsv" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.814221 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.815027 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.870613 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8cwv\" (UniqueName: \"kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.870859 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.870900 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.870921 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdk59\" (UniqueName: \"kubernetes.io/projected/f5cc70d4-9318-4552-825b-9ad495d378b0-kube-api-access-hdk59\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.870992 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.871023 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.875704 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.884368 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.892649 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8cwv\" (UniqueName: \"kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv\") pod \"nova-scheduler-0\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.931022 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974101 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974149 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl297\" (UniqueName: \"kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974227 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974305 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974349 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974399 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.974432 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdk59\" (UniqueName: \"kubernetes.io/projected/f5cc70d4-9318-4552-825b-9ad495d378b0-kube-api-access-hdk59\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.978951 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.989721 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cc70d4-9318-4552-825b-9ad495d378b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:43 crc kubenswrapper[4888]: I1124 00:47:43.993974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdk59\" (UniqueName: \"kubernetes.io/projected/f5cc70d4-9318-4552-825b-9ad495d378b0-kube-api-access-hdk59\") pod \"nova-cell1-conductor-0\" (UID: \"f5cc70d4-9318-4552-825b-9ad495d378b0\") " pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.027279 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.077239 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl297\" (UniqueName: \"kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.077307 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.077365 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.077392 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.083711 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.088109 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.088389 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.101765 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl297\" (UniqueName: \"kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297\") pod \"aodh-0\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.128301 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.269867 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe" path="/var/lib/kubelet/pods/4eee8b8d-893d-4462-8a3e-9dd9e2be6bbe/volumes" Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.544674 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.673074 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 00:47:44 crc kubenswrapper[4888]: I1124 00:47:44.693309 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.557289 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"10f840a2-3864-489a-a694-ac89fa5dadaa","Type":"ContainerStarted","Data":"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.557690 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"10f840a2-3864-489a-a694-ac89fa5dadaa","Type":"ContainerStarted","Data":"213f64ea9be191685a39fa1358268d98756b019c29117c8018aa92c3a954d7b3"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.562380 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f5cc70d4-9318-4552-825b-9ad495d378b0","Type":"ContainerStarted","Data":"7115f56bad8c14df28fcef616ea83e915d0cb68cc4ec84e836757abab2f447f3"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.562426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f5cc70d4-9318-4552-825b-9ad495d378b0","Type":"ContainerStarted","Data":"28c016e86e2279968fb4ab274d0a87c9a21d1a12826f11a621d9196b434bd366"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.563571 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.568174 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerStarted","Data":"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.568204 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerStarted","Data":"efbd3ee6b9772ed3b450e15ac4a3d2c5b456499ecbd2c6fd4a42c3c344c90790"} Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.575149 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.575135045 podStartE2EDuration="2.575135045s" podCreationTimestamp="2025-11-24 00:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:45.571886734 +0000 UTC m=+1368.154570778" watchObservedRunningTime="2025-11-24 00:47:45.575135045 +0000 UTC m=+1368.157819079" Nov 24 00:47:45 crc kubenswrapper[4888]: I1124 00:47:45.598835 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.598800527 podStartE2EDuration="2.598800527s" podCreationTimestamp="2025-11-24 00:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:45.589066124 +0000 UTC m=+1368.171750168" watchObservedRunningTime="2025-11-24 00:47:45.598800527 +0000 UTC m=+1368.181484571" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.442370 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.535943 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs\") pod \"0c013531-f714-4704-bd2b-53136d4b0db9\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.536027 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98jzx\" (UniqueName: \"kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx\") pod \"0c013531-f714-4704-bd2b-53136d4b0db9\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.536083 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data\") pod \"0c013531-f714-4704-bd2b-53136d4b0db9\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.536209 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle\") pod \"0c013531-f714-4704-bd2b-53136d4b0db9\" (UID: \"0c013531-f714-4704-bd2b-53136d4b0db9\") " Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.536528 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs" (OuterVolumeSpecName: "logs") pod "0c013531-f714-4704-bd2b-53136d4b0db9" (UID: "0c013531-f714-4704-bd2b-53136d4b0db9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.537114 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c013531-f714-4704-bd2b-53136d4b0db9-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.541467 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx" (OuterVolumeSpecName: "kube-api-access-98jzx") pod "0c013531-f714-4704-bd2b-53136d4b0db9" (UID: "0c013531-f714-4704-bd2b-53136d4b0db9"). InnerVolumeSpecName "kube-api-access-98jzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.570609 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data" (OuterVolumeSpecName: "config-data") pod "0c013531-f714-4704-bd2b-53136d4b0db9" (UID: "0c013531-f714-4704-bd2b-53136d4b0db9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.582463 4888 generic.go:334] "Generic (PLEG): container finished" podID="0c013531-f714-4704-bd2b-53136d4b0db9" containerID="e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f" exitCode=0 Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.583642 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.583976 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerDied","Data":"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f"} Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.584035 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c013531-f714-4704-bd2b-53136d4b0db9","Type":"ContainerDied","Data":"3059de5e0f8fb591e3bc8f3fda34a0fb12d9888ff60a7428d01a75be49189c4f"} Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.584053 4888 scope.go:117] "RemoveContainer" containerID="e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.591610 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c013531-f714-4704-bd2b-53136d4b0db9" (UID: "0c013531-f714-4704-bd2b-53136d4b0db9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.639328 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98jzx\" (UniqueName: \"kubernetes.io/projected/0c013531-f714-4704-bd2b-53136d4b0db9-kube-api-access-98jzx\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.639369 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.639380 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c013531-f714-4704-bd2b-53136d4b0db9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.656770 4888 scope.go:117] "RemoveContainer" containerID="f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.798766 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.799441 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-central-agent" containerID="cri-o://777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31" gracePeriod=30 Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.800917 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-notification-agent" containerID="cri-o://fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa" gracePeriod=30 Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.800988 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="sg-core" containerID="cri-o://b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835" gracePeriod=30 Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.801176 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" containerID="cri-o://75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1" gracePeriod=30 Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.832888 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.954702 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:46 crc kubenswrapper[4888]: I1124 00:47:46.981963 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.007157 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:47 crc kubenswrapper[4888]: E1124 00:47:47.007675 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-api" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.007700 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-api" Nov 24 00:47:47 crc kubenswrapper[4888]: E1124 00:47:47.007752 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-log" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.007759 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-log" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.007974 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-log" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.007989 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" containerName="nova-api-api" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.009189 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.013193 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.017180 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.150681 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.151051 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.151183 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.151327 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb68q\" (UniqueName: \"kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.253595 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.253898 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.254071 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb68q\" (UniqueName: \"kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.254193 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.254663 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.262557 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.275006 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.276327 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb68q\" (UniqueName: \"kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q\") pod \"nova-api-0\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.368474 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.420395 4888 scope.go:117] "RemoveContainer" containerID="e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f" Nov 24 00:47:47 crc kubenswrapper[4888]: E1124 00:47:47.420880 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f\": container with ID starting with e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f not found: ID does not exist" containerID="e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.420914 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f"} err="failed to get container status \"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f\": rpc error: code = NotFound desc = could not find container \"e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f\": container with ID starting with e806473929267226e0a9668419837a52042608756c2b83a15616bdabd271ad8f not found: ID does not exist" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.420933 4888 scope.go:117] "RemoveContainer" containerID="f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909" Nov 24 00:47:47 crc kubenswrapper[4888]: E1124 00:47:47.421277 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909\": container with ID starting with f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909 not found: ID does not exist" containerID="f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.421299 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909"} err="failed to get container status \"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909\": rpc error: code = NotFound desc = could not find container \"f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909\": container with ID starting with f807fd71e1c0871e7debc6d174096b6a58f9b2025652308f00b846244d99c909 not found: ID does not exist" Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614235 4888 generic.go:334] "Generic (PLEG): container finished" podID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerID="75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1" exitCode=0 Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614518 4888 generic.go:334] "Generic (PLEG): container finished" podID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerID="b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835" exitCode=2 Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614525 4888 generic.go:334] "Generic (PLEG): container finished" podID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerID="777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31" exitCode=0 Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614316 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerDied","Data":"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1"} Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614594 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerDied","Data":"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835"} Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.614610 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerDied","Data":"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31"} Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.781138 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 24 00:47:47 crc kubenswrapper[4888]: I1124 00:47:47.899172 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.258450 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c013531-f714-4704-bd2b-53136d4b0db9" path="/var/lib/kubelet/pods/0c013531-f714-4704-bd2b-53136d4b0db9/volumes" Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.664536 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerStarted","Data":"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c"} Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.673382 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerStarted","Data":"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85"} Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.673438 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerStarted","Data":"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0"} Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.673455 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerStarted","Data":"b778c6cac4d61bedeb151a1d56a38d4f61ddcce175881d89bac99afe06f93567"} Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.699366 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.699344228 podStartE2EDuration="2.699344228s" podCreationTimestamp="2025-11-24 00:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:47:48.690679425 +0000 UTC m=+1371.273363469" watchObservedRunningTime="2025-11-24 00:47:48.699344228 +0000 UTC m=+1371.282028292" Nov 24 00:47:48 crc kubenswrapper[4888]: I1124 00:47:48.931339 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 00:47:49 crc kubenswrapper[4888]: I1124 00:47:49.116109 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 00:47:49 crc kubenswrapper[4888]: I1124 00:47:49.685056 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerStarted","Data":"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d"} Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.412609 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.217:3000/\": dial tcp 10.217.0.217:3000: connect: connection refused" Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.724976 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerStarted","Data":"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d"} Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.725153 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-api" containerID="cri-o://aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3" gracePeriod=30 Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.725614 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-listener" containerID="cri-o://fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d" gracePeriod=30 Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.725679 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-notifier" containerID="cri-o://027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d" gracePeriod=30 Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.725730 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-evaluator" containerID="cri-o://775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c" gracePeriod=30 Nov 24 00:47:51 crc kubenswrapper[4888]: I1124 00:47:51.752283 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.951898842 podStartE2EDuration="8.752267637s" podCreationTimestamp="2025-11-24 00:47:43 +0000 UTC" firstStartedPulling="2025-11-24 00:47:44.737048006 +0000 UTC m=+1367.319732050" lastFinishedPulling="2025-11-24 00:47:50.537416801 +0000 UTC m=+1373.120100845" observedRunningTime="2025-11-24 00:47:51.747213285 +0000 UTC m=+1374.329897329" watchObservedRunningTime="2025-11-24 00:47:51.752267637 +0000 UTC m=+1374.334951681" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.357270 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.477555 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j995z\" (UniqueName: \"kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.477891 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.477987 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.478027 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.478053 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.478074 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.478151 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd\") pod \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\" (UID: \"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69\") " Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.478913 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.480245 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.483146 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts" (OuterVolumeSpecName: "scripts") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.485084 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z" (OuterVolumeSpecName: "kube-api-access-j995z") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "kube-api-access-j995z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.507694 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.577285 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580477 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580509 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580518 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580528 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580537 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j995z\" (UniqueName: \"kubernetes.io/projected/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-kube-api-access-j995z\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.580545 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.600702 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data" (OuterVolumeSpecName: "config-data") pod "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" (UID: "03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.683135 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.743899 4888 generic.go:334] "Generic (PLEG): container finished" podID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerID="fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa" exitCode=0 Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.744011 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerDied","Data":"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa"} Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.744050 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69","Type":"ContainerDied","Data":"a3f441832521a95294c215a6fadf36135949d0d5c19a5001fd3e996ea081117e"} Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.744080 4888 scope.go:117] "RemoveContainer" containerID="75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.744285 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.756924 4888 generic.go:334] "Generic (PLEG): container finished" podID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerID="027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d" exitCode=0 Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.756982 4888 generic.go:334] "Generic (PLEG): container finished" podID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerID="775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c" exitCode=0 Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.757010 4888 generic.go:334] "Generic (PLEG): container finished" podID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerID="aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3" exitCode=0 Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.757022 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerDied","Data":"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d"} Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.757082 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerDied","Data":"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c"} Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.757095 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerDied","Data":"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3"} Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.784116 4888 scope.go:117] "RemoveContainer" containerID="b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.810019 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.827857 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.842796 4888 scope.go:117] "RemoveContainer" containerID="fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.843435 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.844141 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-notification-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844165 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-notification-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.844183 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-central-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844192 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-central-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.844252 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="sg-core" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844262 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="sg-core" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.844273 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844280 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844488 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="sg-core" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844510 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-central-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844523 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="ceilometer-notification-agent" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.844531 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" containerName="proxy-httpd" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.847571 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.857592 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.857778 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.889999 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.891852 4888 scope.go:117] "RemoveContainer" containerID="777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.920558 4888 scope.go:117] "RemoveContainer" containerID="75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.921202 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1\": container with ID starting with 75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1 not found: ID does not exist" containerID="75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.921242 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1"} err="failed to get container status \"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1\": rpc error: code = NotFound desc = could not find container \"75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1\": container with ID starting with 75f04c781770f67e2da6aa5e939e5b846f488db811f5e7d96e3355e1d162b1d1 not found: ID does not exist" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.921286 4888 scope.go:117] "RemoveContainer" containerID="b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.921932 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835\": container with ID starting with b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835 not found: ID does not exist" containerID="b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.921960 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835"} err="failed to get container status \"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835\": rpc error: code = NotFound desc = could not find container \"b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835\": container with ID starting with b6bfb33e15ad0a85b8722d04e265c8b55bef4ec3aaa2af6ee1840c5e5c6c2835 not found: ID does not exist" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.921975 4888 scope.go:117] "RemoveContainer" containerID="fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.922558 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa\": container with ID starting with fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa not found: ID does not exist" containerID="fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.922592 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa"} err="failed to get container status \"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa\": rpc error: code = NotFound desc = could not find container \"fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa\": container with ID starting with fdaf306f35da022543c8fdb5aca820d22a382812b164136af9cc3cfe05b9b7fa not found: ID does not exist" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.922607 4888 scope.go:117] "RemoveContainer" containerID="777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31" Nov 24 00:47:52 crc kubenswrapper[4888]: E1124 00:47:52.922781 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31\": container with ID starting with 777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31 not found: ID does not exist" containerID="777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.922801 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31"} err="failed to get container status \"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31\": rpc error: code = NotFound desc = could not find container \"777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31\": container with ID starting with 777dba3f46edbe9678b5a0c08e8b9a75df7a664484f454c65b2db2f33b520a31 not found: ID does not exist" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.988928 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989012 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989042 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989097 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tttnb\" (UniqueName: \"kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989117 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989210 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:52 crc kubenswrapper[4888]: I1124 00:47:52.989264 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.091901 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.091962 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092009 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092052 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092068 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092112 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tttnb\" (UniqueName: \"kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092126 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092469 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.092781 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.098270 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.098478 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.099858 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.104439 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.108021 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tttnb\" (UniqueName: \"kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb\") pod \"ceilometer-0\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.190610 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.697253 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:47:53 crc kubenswrapper[4888]: W1124 00:47:53.700016 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae7465b5_af42_46cc_a038_78ef166baa21.slice/crio-8ebe4fa0e16ff15d1db1ce3a44c3ffc292667a84c902b29e3c1de0b95b2d7342 WatchSource:0}: Error finding container 8ebe4fa0e16ff15d1db1ce3a44c3ffc292667a84c902b29e3c1de0b95b2d7342: Status 404 returned error can't find the container with id 8ebe4fa0e16ff15d1db1ce3a44c3ffc292667a84c902b29e3c1de0b95b2d7342 Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.774794 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerStarted","Data":"8ebe4fa0e16ff15d1db1ce3a44c3ffc292667a84c902b29e3c1de0b95b2d7342"} Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.932785 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 00:47:53 crc kubenswrapper[4888]: I1124 00:47:53.963544 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 00:47:54 crc kubenswrapper[4888]: I1124 00:47:54.297093 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69" path="/var/lib/kubelet/pods/03cd90d7-a85f-4b7a-b4c5-0e10b4bddd69/volumes" Nov 24 00:47:54 crc kubenswrapper[4888]: I1124 00:47:54.803244 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerStarted","Data":"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108"} Nov 24 00:47:54 crc kubenswrapper[4888]: I1124 00:47:54.856394 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 00:47:55 crc kubenswrapper[4888]: I1124 00:47:55.815611 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerStarted","Data":"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166"} Nov 24 00:47:56 crc kubenswrapper[4888]: I1124 00:47:56.830399 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerStarted","Data":"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0"} Nov 24 00:47:57 crc kubenswrapper[4888]: I1124 00:47:57.368955 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:47:57 crc kubenswrapper[4888]: I1124 00:47:57.369824 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:47:58 crc kubenswrapper[4888]: I1124 00:47:58.451075 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:47:58 crc kubenswrapper[4888]: I1124 00:47:58.451118 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 00:47:58 crc kubenswrapper[4888]: I1124 00:47:58.852002 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerStarted","Data":"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25"} Nov 24 00:47:58 crc kubenswrapper[4888]: I1124 00:47:58.852228 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:47:58 crc kubenswrapper[4888]: I1124 00:47:58.881653 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.822730847 podStartE2EDuration="6.88163494s" podCreationTimestamp="2025-11-24 00:47:52 +0000 UTC" firstStartedPulling="2025-11-24 00:47:53.703101624 +0000 UTC m=+1376.285785678" lastFinishedPulling="2025-11-24 00:47:57.762005717 +0000 UTC m=+1380.344689771" observedRunningTime="2025-11-24 00:47:58.876574168 +0000 UTC m=+1381.459258242" watchObservedRunningTime="2025-11-24 00:47:58.88163494 +0000 UTC m=+1381.464318984" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.188077 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.191190 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.205580 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.284330 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw65q\" (UniqueName: \"kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.284615 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.284766 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.386673 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw65q\" (UniqueName: \"kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.387417 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.388162 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.388507 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.389329 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.407637 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw65q\" (UniqueName: \"kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q\") pod \"certified-operators-qdg5d\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:04 crc kubenswrapper[4888]: I1124 00:48:04.516394 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:05 crc kubenswrapper[4888]: I1124 00:48:05.006399 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:05 crc kubenswrapper[4888]: W1124 00:48:05.012093 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1831ed1_54d7_4829_9523_d6bf67faa243.slice/crio-366e05cac06090f208ec500a0f141816bfef2758b338332c55635e36246d0791 WatchSource:0}: Error finding container 366e05cac06090f208ec500a0f141816bfef2758b338332c55635e36246d0791: Status 404 returned error can't find the container with id 366e05cac06090f208ec500a0f141816bfef2758b338332c55635e36246d0791 Nov 24 00:48:05 crc kubenswrapper[4888]: I1124 00:48:05.964457 4888 generic.go:334] "Generic (PLEG): container finished" podID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerID="7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2" exitCode=0 Nov 24 00:48:05 crc kubenswrapper[4888]: I1124 00:48:05.964517 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerDied","Data":"7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2"} Nov 24 00:48:05 crc kubenswrapper[4888]: I1124 00:48:05.965039 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerStarted","Data":"366e05cac06090f208ec500a0f141816bfef2758b338332c55635e36246d0791"} Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.382043 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.382608 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.384413 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.384524 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.393618 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.393985 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.639870 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.648832 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.652634 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659756 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659825 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kdr6\" (UniqueName: \"kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659872 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659891 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659910 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.659928 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761775 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761849 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kdr6\" (UniqueName: \"kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761898 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761918 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761940 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.761956 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.762855 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.763236 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.765372 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.765955 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.766465 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.789582 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kdr6\" (UniqueName: \"kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6\") pod \"dnsmasq-dns-6b7bbf7cf9-kz8ll\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.976934 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.987138 4888 generic.go:334] "Generic (PLEG): container finished" podID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerID="9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2" exitCode=0 Nov 24 00:48:07 crc kubenswrapper[4888]: I1124 00:48:07.988708 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerDied","Data":"9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2"} Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.563085 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.839462 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.899794 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs\") pod \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.900074 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgqhf\" (UniqueName: \"kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf\") pod \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.900175 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle\") pod \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.900235 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data\") pod \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\" (UID: \"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09\") " Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.900860 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs" (OuterVolumeSpecName: "logs") pod "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" (UID: "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.901001 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.905673 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf" (OuterVolumeSpecName: "kube-api-access-fgqhf") pod "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" (UID: "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09"). InnerVolumeSpecName "kube-api-access-fgqhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.938859 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" (UID: "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:08 crc kubenswrapper[4888]: I1124 00:48:08.953996 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data" (OuterVolumeSpecName: "config-data") pod "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" (UID: "b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.000011 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.002299 4888 generic.go:334] "Generic (PLEG): container finished" podID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerID="efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24" exitCode=137 Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.002345 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerDied","Data":"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.002367 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09","Type":"ContainerDied","Data":"a4094de7a7432a294be324c4d15ce5eeac2f275522fa4a0e449517feb71596d8"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.002383 4888 scope.go:117] "RemoveContainer" containerID="efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.002471 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.007028 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.007051 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.007061 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgqhf\" (UniqueName: \"kubernetes.io/projected/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09-kube-api-access-fgqhf\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.013382 4888 generic.go:334] "Generic (PLEG): container finished" podID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" containerID="8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a" exitCode=137 Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.013440 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39c5c954-3e9f-425f-b250-bf1bdbbb7d84","Type":"ContainerDied","Data":"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.013465 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39c5c954-3e9f-425f-b250-bf1bdbbb7d84","Type":"ContainerDied","Data":"2adf0eef4e0dd78a23419487f9c3541c4fada338a5c8c83e7adf0130ca82b346"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.013510 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.028933 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerStarted","Data":"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.043838 4888 generic.go:334] "Generic (PLEG): container finished" podID="d115b693-6451-46a1-a81e-327eea577d2b" containerID="ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca" exitCode=0 Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.044865 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" event={"ID":"d115b693-6451-46a1-a81e-327eea577d2b","Type":"ContainerDied","Data":"ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.044903 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" event={"ID":"d115b693-6451-46a1-a81e-327eea577d2b","Type":"ContainerStarted","Data":"c21a92cd7079c47491d8e1c60edec8ac73f74992f8722ccd43ada454a84642c4"} Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.074877 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qdg5d" podStartSLOduration=2.53009321 podStartE2EDuration="5.063808418s" podCreationTimestamp="2025-11-24 00:48:04 +0000 UTC" firstStartedPulling="2025-11-24 00:48:05.966696953 +0000 UTC m=+1388.549381007" lastFinishedPulling="2025-11-24 00:48:08.500412171 +0000 UTC m=+1391.083096215" observedRunningTime="2025-11-24 00:48:09.062385528 +0000 UTC m=+1391.645069572" watchObservedRunningTime="2025-11-24 00:48:09.063808418 +0000 UTC m=+1391.646492462" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.107860 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88s2z\" (UniqueName: \"kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z\") pod \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.107945 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data\") pod \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.108063 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle\") pod \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\" (UID: \"39c5c954-3e9f-425f-b250-bf1bdbbb7d84\") " Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.111042 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z" (OuterVolumeSpecName: "kube-api-access-88s2z") pod "39c5c954-3e9f-425f-b250-bf1bdbbb7d84" (UID: "39c5c954-3e9f-425f-b250-bf1bdbbb7d84"). InnerVolumeSpecName "kube-api-access-88s2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.145364 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data" (OuterVolumeSpecName: "config-data") pod "39c5c954-3e9f-425f-b250-bf1bdbbb7d84" (UID: "39c5c954-3e9f-425f-b250-bf1bdbbb7d84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.159738 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39c5c954-3e9f-425f-b250-bf1bdbbb7d84" (UID: "39c5c954-3e9f-425f-b250-bf1bdbbb7d84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.210608 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88s2z\" (UniqueName: \"kubernetes.io/projected/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-kube-api-access-88s2z\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.210636 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.210646 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c5c954-3e9f-425f-b250-bf1bdbbb7d84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.261211 4888 scope.go:117] "RemoveContainer" containerID="e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.286041 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.300682 4888 scope.go:117] "RemoveContainer" containerID="efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24" Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.301150 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24\": container with ID starting with efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24 not found: ID does not exist" containerID="efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.301175 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24"} err="failed to get container status \"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24\": rpc error: code = NotFound desc = could not find container \"efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24\": container with ID starting with efe891d1ae081bf40d873088f729b5721096eb35f8159cceb7395b8d04ef9c24 not found: ID does not exist" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.301193 4888 scope.go:117] "RemoveContainer" containerID="e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3" Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.301536 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3\": container with ID starting with e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3 not found: ID does not exist" containerID="e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.301555 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3"} err="failed to get container status \"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3\": rpc error: code = NotFound desc = could not find container \"e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3\": container with ID starting with e53cc6c0ca4d2402dad1f2fcf8e8e8ca3fbf28f46afeadf6a270cb5eb3ed0eb3 not found: ID does not exist" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.301567 4888 scope.go:117] "RemoveContainer" containerID="8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.308421 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.320041 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.320617 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.320648 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.320664 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-metadata" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.320671 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-metadata" Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.320691 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-log" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.320713 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-log" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.321000 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-metadata" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.321017 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" containerName="nova-metadata-log" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.321029 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.326709 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.330932 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.331195 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.333044 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.344469 4888 scope.go:117] "RemoveContainer" containerID="8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a" Nov 24 00:48:09 crc kubenswrapper[4888]: E1124 00:48:09.353353 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a\": container with ID starting with 8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a not found: ID does not exist" containerID="8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.353408 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a"} err="failed to get container status \"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a\": rpc error: code = NotFound desc = could not find container \"8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a\": container with ID starting with 8858306d25333bd6877d6342a30962b9a7d9097707c1aa058eb6a1e57440890a not found: ID does not exist" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.383506 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.397058 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.411648 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.413000 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.426086 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.426276 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.426313 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.426342 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517183 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517227 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517261 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm75s\" (UniqueName: \"kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517295 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517316 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l75p\" (UniqueName: \"kubernetes.io/projected/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-kube-api-access-2l75p\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517339 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517424 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517471 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.517499 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619056 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619284 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619335 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619380 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619397 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619428 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm75s\" (UniqueName: \"kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619461 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619481 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l75p\" (UniqueName: \"kubernetes.io/projected/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-kube-api-access-2l75p\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619504 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619597 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.619986 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.627039 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.627362 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.629453 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.631425 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.632941 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.633239 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.633352 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.650942 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l75p\" (UniqueName: \"kubernetes.io/projected/a5e4972d-5b76-4f46-a9fa-3a4425534ad8-kube-api-access-2l75p\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5e4972d-5b76-4f46-a9fa-3a4425534ad8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.659884 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm75s\" (UniqueName: \"kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s\") pod \"nova-metadata-0\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.664877 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:09 crc kubenswrapper[4888]: I1124 00:48:09.747496 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.059974 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" event={"ID":"d115b693-6451-46a1-a81e-327eea577d2b","Type":"ContainerStarted","Data":"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4"} Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.060337 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.090337 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" podStartSLOduration=3.090313666 podStartE2EDuration="3.090313666s" podCreationTimestamp="2025-11-24 00:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:10.079276787 +0000 UTC m=+1392.661960831" watchObservedRunningTime="2025-11-24 00:48:10.090313666 +0000 UTC m=+1392.672997710" Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.164442 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.273959 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c5c954-3e9f-425f-b250-bf1bdbbb7d84" path="/var/lib/kubelet/pods/39c5c954-3e9f-425f-b250-bf1bdbbb7d84/volumes" Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.274738 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09" path="/var/lib/kubelet/pods/b7fd2efb-ff6d-4b7d-82f6-04fe1f501e09/volumes" Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.292377 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.292591 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-log" containerID="cri-o://a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.292723 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-api" containerID="cri-o://2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.322261 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.449616 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.450284 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-central-agent" containerID="cri-o://c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.451059 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="proxy-httpd" containerID="cri-o://a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.451113 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="sg-core" containerID="cri-o://7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.451145 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-notification-agent" containerID="cri-o://12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166" gracePeriod=30 Nov 24 00:48:10 crc kubenswrapper[4888]: I1124 00:48:10.478878 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.076088 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5e4972d-5b76-4f46-a9fa-3a4425534ad8","Type":"ContainerStarted","Data":"86051b492bba95260aa10bc164886ed14f84a5075f02aa46ab8b142c6ec6cdf2"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.076474 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5e4972d-5b76-4f46-a9fa-3a4425534ad8","Type":"ContainerStarted","Data":"dce12138fe9301b9740c5b6f042b209491e6182209c657068492ffdff69b881f"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.078247 4888 generic.go:334] "Generic (PLEG): container finished" podID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerID="a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0" exitCode=143 Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.078315 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerDied","Data":"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080843 4888 generic.go:334] "Generic (PLEG): container finished" podID="ae7465b5-af42-46cc-a038-78ef166baa21" containerID="a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25" exitCode=0 Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080876 4888 generic.go:334] "Generic (PLEG): container finished" podID="ae7465b5-af42-46cc-a038-78ef166baa21" containerID="7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0" exitCode=2 Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080887 4888 generic.go:334] "Generic (PLEG): container finished" podID="ae7465b5-af42-46cc-a038-78ef166baa21" containerID="c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108" exitCode=0 Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080909 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerDied","Data":"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080940 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerDied","Data":"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.080957 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerDied","Data":"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.082974 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerStarted","Data":"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.083015 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerStarted","Data":"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.083050 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerStarted","Data":"cc71647271ff883ba70eef8ea7af797bf9ee053198de894e06267a80b8ebc655"} Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.117951 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.117931164 podStartE2EDuration="2.117931164s" podCreationTimestamp="2025-11-24 00:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:11.112758089 +0000 UTC m=+1393.695442133" watchObservedRunningTime="2025-11-24 00:48:11.117931164 +0000 UTC m=+1393.700615208" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.121070 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.121057051 podStartE2EDuration="2.121057051s" podCreationTimestamp="2025-11-24 00:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:11.096464903 +0000 UTC m=+1393.679148947" watchObservedRunningTime="2025-11-24 00:48:11.121057051 +0000 UTC m=+1393.703741095" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.728724 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.867924 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.868735 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.868789 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tttnb\" (UniqueName: \"kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.868894 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.868949 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.868974 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.869007 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data\") pod \"ae7465b5-af42-46cc-a038-78ef166baa21\" (UID: \"ae7465b5-af42-46cc-a038-78ef166baa21\") " Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.869418 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.869613 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.874239 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts" (OuterVolumeSpecName: "scripts") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.877240 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb" (OuterVolumeSpecName: "kube-api-access-tttnb") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "kube-api-access-tttnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.914076 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.961757 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971764 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tttnb\" (UniqueName: \"kubernetes.io/projected/ae7465b5-af42-46cc-a038-78ef166baa21-kube-api-access-tttnb\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971789 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971800 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae7465b5-af42-46cc-a038-78ef166baa21-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971823 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971832 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.971841 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:11 crc kubenswrapper[4888]: I1124 00:48:11.998706 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data" (OuterVolumeSpecName: "config-data") pod "ae7465b5-af42-46cc-a038-78ef166baa21" (UID: "ae7465b5-af42-46cc-a038-78ef166baa21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.073442 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae7465b5-af42-46cc-a038-78ef166baa21-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.093467 4888 generic.go:334] "Generic (PLEG): container finished" podID="ae7465b5-af42-46cc-a038-78ef166baa21" containerID="12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166" exitCode=0 Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.093507 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.093528 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerDied","Data":"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166"} Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.094343 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae7465b5-af42-46cc-a038-78ef166baa21","Type":"ContainerDied","Data":"8ebe4fa0e16ff15d1db1ce3a44c3ffc292667a84c902b29e3c1de0b95b2d7342"} Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.094365 4888 scope.go:117] "RemoveContainer" containerID="a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.127119 4888 scope.go:117] "RemoveContainer" containerID="7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.130797 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.145904 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.153413 4888 scope.go:117] "RemoveContainer" containerID="12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.158314 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.158764 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="proxy-httpd" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.158780 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="proxy-httpd" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.158791 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-central-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.158798 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-central-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.158836 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-notification-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.158842 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-notification-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.158858 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="sg-core" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.158863 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="sg-core" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.159237 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-central-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.159259 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="sg-core" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.159270 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="proxy-httpd" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.159284 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" containerName="ceilometer-notification-agent" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.161198 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.169342 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.169431 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.169511 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.173858 4888 scope.go:117] "RemoveContainer" containerID="c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.202156 4888 scope.go:117] "RemoveContainer" containerID="a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.202603 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25\": container with ID starting with a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25 not found: ID does not exist" containerID="a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.202630 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25"} err="failed to get container status \"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25\": rpc error: code = NotFound desc = could not find container \"a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25\": container with ID starting with a405c20b2de026bc2f0e2f412586f1f9b6584600b63a5e2595c22fc65faf5d25 not found: ID does not exist" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.202649 4888 scope.go:117] "RemoveContainer" containerID="7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.203005 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0\": container with ID starting with 7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0 not found: ID does not exist" containerID="7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.203041 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0"} err="failed to get container status \"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0\": rpc error: code = NotFound desc = could not find container \"7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0\": container with ID starting with 7c73f2abd867f438882639b7801979e89cb2616a63c7b268ba554d366ee64da0 not found: ID does not exist" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.203067 4888 scope.go:117] "RemoveContainer" containerID="12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.203329 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166\": container with ID starting with 12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166 not found: ID does not exist" containerID="12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.203350 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166"} err="failed to get container status \"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166\": rpc error: code = NotFound desc = could not find container \"12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166\": container with ID starting with 12fc2b0a2f73e4b537b41927caa1392a429c002bbc949a213bb469f449898166 not found: ID does not exist" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.203364 4888 scope.go:117] "RemoveContainer" containerID="c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108" Nov 24 00:48:12 crc kubenswrapper[4888]: E1124 00:48:12.203626 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108\": container with ID starting with c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108 not found: ID does not exist" containerID="c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.203644 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108"} err="failed to get container status \"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108\": rpc error: code = NotFound desc = could not find container \"c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108\": container with ID starting with c39739740d53b555e8cdf3cb63fae7d692f46d097efa961f9a85f2b491ef4108 not found: ID does not exist" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.271730 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae7465b5-af42-46cc-a038-78ef166baa21" path="/var/lib/kubelet/pods/ae7465b5-af42-46cc-a038-78ef166baa21/volumes" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277042 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277177 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277198 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tzlf\" (UniqueName: \"kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277226 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277256 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277430 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.277803 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.379691 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.379960 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.380056 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.380103 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.380161 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tzlf\" (UniqueName: \"kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.380178 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.380200 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.381298 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.381450 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.384214 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.384840 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.391149 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.391339 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.400608 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tzlf\" (UniqueName: \"kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf\") pod \"ceilometer-0\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.481627 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:12 crc kubenswrapper[4888]: I1124 00:48:12.492129 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:13 crc kubenswrapper[4888]: I1124 00:48:13.503959 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.117979 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.177211 4888 generic.go:334] "Generic (PLEG): container finished" podID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerID="2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85" exitCode=0 Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.177269 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerDied","Data":"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85"} Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.177319 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b9a686e6-4b79-4ad5-bf25-e38106503d56","Type":"ContainerDied","Data":"b778c6cac4d61bedeb151a1d56a38d4f61ddcce175881d89bac99afe06f93567"} Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.177341 4888 scope.go:117] "RemoveContainer" containerID="2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.177468 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.184258 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerStarted","Data":"3de6f6d8c03953522077757992adea46fa69815e796cdab4078b59d69d0b411d"} Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.209858 4888 scope.go:117] "RemoveContainer" containerID="a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.219643 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle\") pod \"b9a686e6-4b79-4ad5-bf25-e38106503d56\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.219697 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb68q\" (UniqueName: \"kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q\") pod \"b9a686e6-4b79-4ad5-bf25-e38106503d56\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.219857 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data\") pod \"b9a686e6-4b79-4ad5-bf25-e38106503d56\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.219983 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs\") pod \"b9a686e6-4b79-4ad5-bf25-e38106503d56\" (UID: \"b9a686e6-4b79-4ad5-bf25-e38106503d56\") " Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.220778 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs" (OuterVolumeSpecName: "logs") pod "b9a686e6-4b79-4ad5-bf25-e38106503d56" (UID: "b9a686e6-4b79-4ad5-bf25-e38106503d56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.233110 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q" (OuterVolumeSpecName: "kube-api-access-zb68q") pod "b9a686e6-4b79-4ad5-bf25-e38106503d56" (UID: "b9a686e6-4b79-4ad5-bf25-e38106503d56"). InnerVolumeSpecName "kube-api-access-zb68q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.236655 4888 scope.go:117] "RemoveContainer" containerID="2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85" Nov 24 00:48:14 crc kubenswrapper[4888]: E1124 00:48:14.240618 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85\": container with ID starting with 2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85 not found: ID does not exist" containerID="2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.240660 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85"} err="failed to get container status \"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85\": rpc error: code = NotFound desc = could not find container \"2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85\": container with ID starting with 2d5bb32472f7b555f7475d4ebbc5bf19cd0fff7f98b1a097187f54114507ad85 not found: ID does not exist" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.240691 4888 scope.go:117] "RemoveContainer" containerID="a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0" Nov 24 00:48:14 crc kubenswrapper[4888]: E1124 00:48:14.241457 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0\": container with ID starting with a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0 not found: ID does not exist" containerID="a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.241590 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0"} err="failed to get container status \"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0\": rpc error: code = NotFound desc = could not find container \"a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0\": container with ID starting with a0622c295dc2b5141ce73fce460051edaf4c7dc792e9faf8acfdae8f32c777f0 not found: ID does not exist" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.280976 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data" (OuterVolumeSpecName: "config-data") pod "b9a686e6-4b79-4ad5-bf25-e38106503d56" (UID: "b9a686e6-4b79-4ad5-bf25-e38106503d56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.284002 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9a686e6-4b79-4ad5-bf25-e38106503d56" (UID: "b9a686e6-4b79-4ad5-bf25-e38106503d56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.323116 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.323150 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a686e6-4b79-4ad5-bf25-e38106503d56-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.323164 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a686e6-4b79-4ad5-bf25-e38106503d56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.323176 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb68q\" (UniqueName: \"kubernetes.io/projected/b9a686e6-4b79-4ad5-bf25-e38106503d56-kube-api-access-zb68q\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.517041 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.517421 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.524879 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.547886 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.559859 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:14 crc kubenswrapper[4888]: E1124 00:48:14.560390 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-api" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.560412 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-api" Nov 24 00:48:14 crc kubenswrapper[4888]: E1124 00:48:14.560445 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-log" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.560453 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-log" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.560760 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-log" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.560779 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" containerName="nova-api-api" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.562353 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.564375 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.564944 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.568698 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.574344 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.590747 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.628947 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r46jc\" (UniqueName: \"kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.629032 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.629097 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.629196 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.629382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.629427 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.665404 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.665516 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.730997 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731139 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731166 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731194 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r46jc\" (UniqueName: \"kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731250 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731268 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.731903 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.734790 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.734871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.735116 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.741289 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.747995 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.760334 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r46jc\" (UniqueName: \"kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc\") pod \"nova-api-0\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " pod="openstack/nova-api-0" Nov 24 00:48:14 crc kubenswrapper[4888]: I1124 00:48:14.890691 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:15 crc kubenswrapper[4888]: I1124 00:48:15.201448 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerStarted","Data":"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de"} Nov 24 00:48:15 crc kubenswrapper[4888]: I1124 00:48:15.201495 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerStarted","Data":"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f"} Nov 24 00:48:15 crc kubenswrapper[4888]: I1124 00:48:15.249774 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:15 crc kubenswrapper[4888]: I1124 00:48:15.314773 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:15 crc kubenswrapper[4888]: I1124 00:48:15.448364 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.212105 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerStarted","Data":"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70"} Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.212365 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerStarted","Data":"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0"} Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.212380 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerStarted","Data":"0b6fce4312ab7eceedcc5e8fc76eb34f05bae94d480e1814fffc80ad62f689f4"} Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.217281 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerStarted","Data":"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba"} Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.249748 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2497270719999998 podStartE2EDuration="2.249727072s" podCreationTimestamp="2025-11-24 00:48:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:16.240267678 +0000 UTC m=+1398.822951762" watchObservedRunningTime="2025-11-24 00:48:16.249727072 +0000 UTC m=+1398.832411126" Nov 24 00:48:16 crc kubenswrapper[4888]: I1124 00:48:16.262976 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a686e6-4b79-4ad5-bf25-e38106503d56" path="/var/lib/kubelet/pods/b9a686e6-4b79-4ad5-bf25-e38106503d56/volumes" Nov 24 00:48:17 crc kubenswrapper[4888]: I1124 00:48:17.242260 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qdg5d" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="registry-server" containerID="cri-o://d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098" gracePeriod=2 Nov 24 00:48:17 crc kubenswrapper[4888]: I1124 00:48:17.885403 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:17 crc kubenswrapper[4888]: I1124 00:48:17.978863 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.009091 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content\") pod \"c1831ed1-54d7-4829-9523-d6bf67faa243\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.009328 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw65q\" (UniqueName: \"kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q\") pod \"c1831ed1-54d7-4829-9523-d6bf67faa243\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.009402 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities\") pod \"c1831ed1-54d7-4829-9523-d6bf67faa243\" (UID: \"c1831ed1-54d7-4829-9523-d6bf67faa243\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.010472 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities" (OuterVolumeSpecName: "utilities") pod "c1831ed1-54d7-4829-9523-d6bf67faa243" (UID: "c1831ed1-54d7-4829-9523-d6bf67faa243"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.015900 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q" (OuterVolumeSpecName: "kube-api-access-kw65q") pod "c1831ed1-54d7-4829-9523-d6bf67faa243" (UID: "c1831ed1-54d7-4829-9523-d6bf67faa243"). InnerVolumeSpecName "kube-api-access-kw65q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.046163 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.046384 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="dnsmasq-dns" containerID="cri-o://01082314d83ea4e6e600094b33ddf8e22f486f333fba141c8b3cb9c8676bc453" gracePeriod=10 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.090589 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1831ed1-54d7-4829-9523-d6bf67faa243" (UID: "c1831ed1-54d7-4829-9523-d6bf67faa243"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.112427 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.112461 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw65q\" (UniqueName: \"kubernetes.io/projected/c1831ed1-54d7-4829-9523-d6bf67faa243-kube-api-access-kw65q\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.112471 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1831ed1-54d7-4829-9523-d6bf67faa243-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.257391 4888 generic.go:334] "Generic (PLEG): container finished" podID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerID="01082314d83ea4e6e600094b33ddf8e22f486f333fba141c8b3cb9c8676bc453" exitCode=0 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.261693 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" event={"ID":"53a42b6c-9404-41d4-9d05-5c30e11fece7","Type":"ContainerDied","Data":"01082314d83ea4e6e600094b33ddf8e22f486f333fba141c8b3cb9c8676bc453"} Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.269942 4888 generic.go:334] "Generic (PLEG): container finished" podID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerID="d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098" exitCode=0 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.270019 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerDied","Data":"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098"} Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.270052 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdg5d" event={"ID":"c1831ed1-54d7-4829-9523-d6bf67faa243","Type":"ContainerDied","Data":"366e05cac06090f208ec500a0f141816bfef2758b338332c55635e36246d0791"} Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.270074 4888 scope.go:117] "RemoveContainer" containerID="d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.270236 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdg5d" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.286117 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerStarted","Data":"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa"} Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.286312 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-central-agent" containerID="cri-o://cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f" gracePeriod=30 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.286596 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.287010 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="proxy-httpd" containerID="cri-o://093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa" gracePeriod=30 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.287074 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="sg-core" containerID="cri-o://89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba" gracePeriod=30 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.287122 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-notification-agent" containerID="cri-o://a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de" gracePeriod=30 Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.340963 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.731938235 podStartE2EDuration="6.340943026s" podCreationTimestamp="2025-11-24 00:48:12 +0000 UTC" firstStartedPulling="2025-11-24 00:48:13.518734826 +0000 UTC m=+1396.101418870" lastFinishedPulling="2025-11-24 00:48:17.127739607 +0000 UTC m=+1399.710423661" observedRunningTime="2025-11-24 00:48:18.316866902 +0000 UTC m=+1400.899550956" watchObservedRunningTime="2025-11-24 00:48:18.340943026 +0000 UTC m=+1400.923627070" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.350575 4888 scope.go:117] "RemoveContainer" containerID="9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.355730 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.364251 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qdg5d"] Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.377336 4888 scope.go:117] "RemoveContainer" containerID="7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.538917 4888 scope.go:117] "RemoveContainer" containerID="d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098" Nov 24 00:48:18 crc kubenswrapper[4888]: E1124 00:48:18.540014 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098\": container with ID starting with d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098 not found: ID does not exist" containerID="d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.540051 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098"} err="failed to get container status \"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098\": rpc error: code = NotFound desc = could not find container \"d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098\": container with ID starting with d311e94e16afb62f166ae6234167b9edb5acc5203150b54e5bc7b4e4b1432098 not found: ID does not exist" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.540081 4888 scope.go:117] "RemoveContainer" containerID="9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2" Nov 24 00:48:18 crc kubenswrapper[4888]: E1124 00:48:18.540313 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2\": container with ID starting with 9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2 not found: ID does not exist" containerID="9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.540334 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2"} err="failed to get container status \"9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2\": rpc error: code = NotFound desc = could not find container \"9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2\": container with ID starting with 9489b959265556e05ea4b36db06c3a2be6c1d6b23c9017a0cb67927e14d7b1a2 not found: ID does not exist" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.540359 4888 scope.go:117] "RemoveContainer" containerID="7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2" Nov 24 00:48:18 crc kubenswrapper[4888]: E1124 00:48:18.542291 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2\": container with ID starting with 7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2 not found: ID does not exist" containerID="7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.542319 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2"} err="failed to get container status \"7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2\": rpc error: code = NotFound desc = could not find container \"7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2\": container with ID starting with 7083da52f3007b803a6dbcb976918ab94452f4b6113b3a9a1be63a0c6ed7a8b2 not found: ID does not exist" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.596588 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623305 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623387 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f65nm\" (UniqueName: \"kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623493 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623700 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623727 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.623936 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb\") pod \"53a42b6c-9404-41d4-9d05-5c30e11fece7\" (UID: \"53a42b6c-9404-41d4-9d05-5c30e11fece7\") " Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.644056 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm" (OuterVolumeSpecName: "kube-api-access-f65nm") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "kube-api-access-f65nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.694390 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.703270 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.709515 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config" (OuterVolumeSpecName: "config") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.714126 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.727427 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.727453 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.727462 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.727473 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f65nm\" (UniqueName: \"kubernetes.io/projected/53a42b6c-9404-41d4-9d05-5c30e11fece7-kube-api-access-f65nm\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.727483 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.739622 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53a42b6c-9404-41d4-9d05-5c30e11fece7" (UID: "53a42b6c-9404-41d4-9d05-5c30e11fece7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:48:18 crc kubenswrapper[4888]: I1124 00:48:18.829772 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53a42b6c-9404-41d4-9d05-5c30e11fece7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.300855 4888 generic.go:334] "Generic (PLEG): container finished" podID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerID="093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa" exitCode=0 Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.301215 4888 generic.go:334] "Generic (PLEG): container finished" podID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerID="89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba" exitCode=2 Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.301230 4888 generic.go:334] "Generic (PLEG): container finished" podID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerID="a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de" exitCode=0 Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.300944 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerDied","Data":"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa"} Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.301297 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerDied","Data":"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba"} Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.301313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerDied","Data":"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de"} Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.303246 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" event={"ID":"53a42b6c-9404-41d4-9d05-5c30e11fece7","Type":"ContainerDied","Data":"99fc1bf87533e93afabd5302abbf3fe0497abcd4ce08b107b3624495b01929f5"} Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.303279 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-cwd26" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.303309 4888 scope.go:117] "RemoveContainer" containerID="01082314d83ea4e6e600094b33ddf8e22f486f333fba141c8b3cb9c8676bc453" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.378548 4888 scope.go:117] "RemoveContainer" containerID="baa1f270019b3951eff8163acdaeba4ba62e88a91db8323106434568c5ffa895" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.394085 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.404762 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-cwd26"] Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.665465 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.665522 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.747734 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.759365 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.775467 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.853986 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854427 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854585 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tzlf\" (UniqueName: \"kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854641 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854718 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854781 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.854875 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts\") pod \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\" (UID: \"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7\") " Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.855063 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.855452 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.855820 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.855874 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.860140 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf" (OuterVolumeSpecName: "kube-api-access-4tzlf") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "kube-api-access-4tzlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.860887 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts" (OuterVolumeSpecName: "scripts") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.901041 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.947548 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.957701 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.957741 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.957756 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.957769 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tzlf\" (UniqueName: \"kubernetes.io/projected/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-kube-api-access-4tzlf\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:19 crc kubenswrapper[4888]: I1124 00:48:19.985710 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data" (OuterVolumeSpecName: "config-data") pod "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" (UID: "5520ebd8-3d7c-45a9-899e-4f9f0b314fa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.059814 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.271262 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" path="/var/lib/kubelet/pods/53a42b6c-9404-41d4-9d05-5c30e11fece7/volumes" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.272074 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" path="/var/lib/kubelet/pods/c1831ed1-54d7-4829-9523-d6bf67faa243/volumes" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.328262 4888 generic.go:334] "Generic (PLEG): container finished" podID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerID="cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f" exitCode=0 Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.328379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerDied","Data":"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f"} Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.328432 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5520ebd8-3d7c-45a9-899e-4f9f0b314fa7","Type":"ContainerDied","Data":"3de6f6d8c03953522077757992adea46fa69815e796cdab4078b59d69d0b411d"} Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.328467 4888 scope.go:117] "RemoveContainer" containerID="093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.328686 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.380403 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.380943 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.385788 4888 scope.go:117] "RemoveContainer" containerID="89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.395286 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.408428 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409354 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="proxy-httpd" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409391 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="proxy-httpd" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409428 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="extract-content" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409441 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="extract-content" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409480 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="dnsmasq-dns" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409497 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="dnsmasq-dns" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409518 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="registry-server" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409530 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="registry-server" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409559 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="extract-utilities" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409571 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="extract-utilities" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409616 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="sg-core" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409629 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="sg-core" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409652 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="init" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409664 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="init" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409689 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-central-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409701 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-central-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.409727 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-notification-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.409757 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-notification-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410256 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-notification-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410312 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a42b6c-9404-41d4-9d05-5c30e11fece7" containerName="dnsmasq-dns" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410333 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="ceilometer-central-agent" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410358 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1831ed1-54d7-4829-9523-d6bf67faa243" containerName="registry-server" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410402 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="sg-core" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.410430 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" containerName="proxy-httpd" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.417172 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.431227 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.431618 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.438273 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.452433 4888 scope.go:117] "RemoveContainer" containerID="a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467460 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467576 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dwwg\" (UniqueName: \"kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467608 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467634 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467663 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467700 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.467736 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.505027 4888 scope.go:117] "RemoveContainer" containerID="cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.551816 4888 scope.go:117] "RemoveContainer" containerID="093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.555106 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa\": container with ID starting with 093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa not found: ID does not exist" containerID="093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.555147 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa"} err="failed to get container status \"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa\": rpc error: code = NotFound desc = could not find container \"093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa\": container with ID starting with 093f69796a1f6d2c3ee70cde536e702629ca6997060828aae5a7411b863707fa not found: ID does not exist" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.555173 4888 scope.go:117] "RemoveContainer" containerID="89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.555685 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba\": container with ID starting with 89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba not found: ID does not exist" containerID="89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.555719 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba"} err="failed to get container status \"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba\": rpc error: code = NotFound desc = could not find container \"89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba\": container with ID starting with 89143dfc0d76e2599e29f3cec998ba153413d5f1442fa472ec8d91c5df42e3ba not found: ID does not exist" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.555743 4888 scope.go:117] "RemoveContainer" containerID="a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.557641 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de\": container with ID starting with a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de not found: ID does not exist" containerID="a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.557668 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de"} err="failed to get container status \"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de\": rpc error: code = NotFound desc = could not find container \"a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de\": container with ID starting with a00a08a537207155ba5f9a2c504d48edad7e071236e4f1760fcff1608a8cf5de not found: ID does not exist" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.557683 4888 scope.go:117] "RemoveContainer" containerID="cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f" Nov 24 00:48:20 crc kubenswrapper[4888]: E1124 00:48:20.559302 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f\": container with ID starting with cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f not found: ID does not exist" containerID="cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.559324 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f"} err="failed to get container status \"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f\": rpc error: code = NotFound desc = could not find container \"cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f\": container with ID starting with cd1182f61fbc254384ef7c9a9007eca0ea190b7264af2b20ba2afb5f7feebe5f not found: ID does not exist" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570150 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570291 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dwwg\" (UniqueName: \"kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570333 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570366 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570409 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570457 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.570513 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.571142 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.571409 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.583245 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kmm2t"] Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.586475 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.589983 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.590002 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.590493 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.590526 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dwwg\" (UniqueName: \"kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.591534 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.607498 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.610583 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts\") pod \"ceilometer-0\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.616098 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmm2t"] Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.672666 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.672797 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvftw\" (UniqueName: \"kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.672864 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.673143 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.675985 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.235:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.676026 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.235:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.746891 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.775646 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.775765 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.775846 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvftw\" (UniqueName: \"kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.775871 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.784392 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.784473 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.785713 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:20 crc kubenswrapper[4888]: I1124 00:48:20.807152 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvftw\" (UniqueName: \"kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw\") pod \"nova-cell1-cell-mapping-kmm2t\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:21 crc kubenswrapper[4888]: I1124 00:48:21.011649 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:21 crc kubenswrapper[4888]: I1124 00:48:21.231513 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:21 crc kubenswrapper[4888]: I1124 00:48:21.346290 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerStarted","Data":"66098b326642ee723ad9853f55baf3d2bd09cddc0663303c5760e47513fe7bd7"} Nov 24 00:48:21 crc kubenswrapper[4888]: I1124 00:48:21.484726 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmm2t"] Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.111137 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.257597 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5520ebd8-3d7c-45a9-899e-4f9f0b314fa7" path="/var/lib/kubelet/pods/5520ebd8-3d7c-45a9-899e-4f9f0b314fa7/volumes" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.307903 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle\") pod \"8122b580-0ac5-42f9-826d-639f67b00ef4\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.308297 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data\") pod \"8122b580-0ac5-42f9-826d-639f67b00ef4\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.308344 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts\") pod \"8122b580-0ac5-42f9-826d-639f67b00ef4\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.308403 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl297\" (UniqueName: \"kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297\") pod \"8122b580-0ac5-42f9-826d-639f67b00ef4\" (UID: \"8122b580-0ac5-42f9-826d-639f67b00ef4\") " Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.314194 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts" (OuterVolumeSpecName: "scripts") pod "8122b580-0ac5-42f9-826d-639f67b00ef4" (UID: "8122b580-0ac5-42f9-826d-639f67b00ef4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.319986 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297" (OuterVolumeSpecName: "kube-api-access-jl297") pod "8122b580-0ac5-42f9-826d-639f67b00ef4" (UID: "8122b580-0ac5-42f9-826d-639f67b00ef4"). InnerVolumeSpecName "kube-api-access-jl297". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.378935 4888 generic.go:334] "Generic (PLEG): container finished" podID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerID="fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d" exitCode=137 Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.379004 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerDied","Data":"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d"} Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.379035 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8122b580-0ac5-42f9-826d-639f67b00ef4","Type":"ContainerDied","Data":"efbd3ee6b9772ed3b450e15ac4a3d2c5b456499ecbd2c6fd4a42c3c344c90790"} Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.379170 4888 scope.go:117] "RemoveContainer" containerID="fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.379341 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.387511 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerStarted","Data":"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99"} Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.398831 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmm2t" event={"ID":"cab6f695-53c2-43fe-bef0-fea107068d30","Type":"ContainerStarted","Data":"f03f887341c9e9500f0819e7e02b31aa263a5c14ba443321635f6bc69429f1d3"} Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.398910 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmm2t" event={"ID":"cab6f695-53c2-43fe-bef0-fea107068d30","Type":"ContainerStarted","Data":"39fd31ecbc94fb1796ebc1d7a20aa75b2cf0e79199ad4da8ee3384ef3bc479d3"} Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.411425 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl297\" (UniqueName: \"kubernetes.io/projected/8122b580-0ac5-42f9-826d-639f67b00ef4-kube-api-access-jl297\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.411462 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.414326 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kmm2t" podStartSLOduration=2.414314583 podStartE2EDuration="2.414314583s" podCreationTimestamp="2025-11-24 00:48:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:22.41242045 +0000 UTC m=+1404.995104494" watchObservedRunningTime="2025-11-24 00:48:22.414314583 +0000 UTC m=+1404.996998627" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.461978 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8122b580-0ac5-42f9-826d-639f67b00ef4" (UID: "8122b580-0ac5-42f9-826d-639f67b00ef4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.482535 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data" (OuterVolumeSpecName: "config-data") pod "8122b580-0ac5-42f9-826d-639f67b00ef4" (UID: "8122b580-0ac5-42f9-826d-639f67b00ef4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.516443 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.516474 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8122b580-0ac5-42f9-826d-639f67b00ef4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.547141 4888 scope.go:117] "RemoveContainer" containerID="027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.570637 4888 scope.go:117] "RemoveContainer" containerID="775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.593663 4888 scope.go:117] "RemoveContainer" containerID="aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.628165 4888 scope.go:117] "RemoveContainer" containerID="fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.628909 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d\": container with ID starting with fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d not found: ID does not exist" containerID="fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.628952 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d"} err="failed to get container status \"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d\": rpc error: code = NotFound desc = could not find container \"fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d\": container with ID starting with fe93d1fbcf2a55b1f9ca93d15a274a0cd9965402a1b6fc6f858c07af8967ec7d not found: ID does not exist" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.628987 4888 scope.go:117] "RemoveContainer" containerID="027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.629978 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d\": container with ID starting with 027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d not found: ID does not exist" containerID="027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.630010 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d"} err="failed to get container status \"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d\": rpc error: code = NotFound desc = could not find container \"027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d\": container with ID starting with 027733e3f162810a38ff33d8d793ad0e3a32e0ccf0929f39f90dca8d7db6549d not found: ID does not exist" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.630029 4888 scope.go:117] "RemoveContainer" containerID="775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.630261 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c\": container with ID starting with 775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c not found: ID does not exist" containerID="775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.630286 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c"} err="failed to get container status \"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c\": rpc error: code = NotFound desc = could not find container \"775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c\": container with ID starting with 775f8807509557a0f6e3479620654d1f1d36280609316425301edb676b430e0c not found: ID does not exist" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.630331 4888 scope.go:117] "RemoveContainer" containerID="aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.630859 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3\": container with ID starting with aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3 not found: ID does not exist" containerID="aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.630889 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3"} err="failed to get container status \"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3\": rpc error: code = NotFound desc = could not find container \"aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3\": container with ID starting with aa4bdf185514f797176a1c46c50c7872d4270776c35dada1f0234f89f914e9c3 not found: ID does not exist" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.725976 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.736844 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.749505 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.750189 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-listener" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750221 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-listener" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.750246 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-evaluator" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750263 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-evaluator" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.750324 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-notifier" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750342 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-notifier" Nov 24 00:48:22 crc kubenswrapper[4888]: E1124 00:48:22.750412 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-api" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750436 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-api" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750803 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-api" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750859 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-notifier" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750918 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-evaluator" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.750957 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" containerName="aodh-listener" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.754509 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.756620 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.756724 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.758081 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.758931 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-t9vsv" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.759640 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.763265 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822394 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822470 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ndn7\" (UniqueName: \"kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822505 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822587 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822645 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.822668 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.923965 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.924020 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.924103 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.924137 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ndn7\" (UniqueName: \"kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.924164 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.924189 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.930458 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.933424 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.933769 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.933934 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.935913 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:22 crc kubenswrapper[4888]: I1124 00:48:22.948014 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ndn7\" (UniqueName: \"kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7\") pod \"aodh-0\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " pod="openstack/aodh-0" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.082769 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.408892 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerStarted","Data":"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb"} Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.409248 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerStarted","Data":"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0"} Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.503975 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.505900 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.530873 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.584847 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:48:23 crc kubenswrapper[4888]: W1124 00:48:23.586395 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5645bf92_6527_49a8_90e7_721ae2eaedf6.slice/crio-2e9fa882b485b4cb5309fe0a2a2d2db25fed6080c8ce767b94dde09a4f33e1a9 WatchSource:0}: Error finding container 2e9fa882b485b4cb5309fe0a2a2d2db25fed6080c8ce767b94dde09a4f33e1a9: Status 404 returned error can't find the container with id 2e9fa882b485b4cb5309fe0a2a2d2db25fed6080c8ce767b94dde09a4f33e1a9 Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.643504 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.643566 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dltjs\" (UniqueName: \"kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.643640 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.745544 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.746027 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.746365 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.746417 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dltjs\" (UniqueName: \"kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.746730 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.776207 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dltjs\" (UniqueName: \"kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs\") pod \"redhat-operators-hppc9\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:23 crc kubenswrapper[4888]: I1124 00:48:23.826949 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.261179 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8122b580-0ac5-42f9-826d-639f67b00ef4" path="/var/lib/kubelet/pods/8122b580-0ac5-42f9-826d-639f67b00ef4/volumes" Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.386765 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:24 crc kubenswrapper[4888]: W1124 00:48:24.392076 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde516985_13b9_49af_a46a_b072e62d6c25.slice/crio-6ed438fb9c37a9314259d577aca91c6fd13765e5d6929037a3ef35586d0b0296 WatchSource:0}: Error finding container 6ed438fb9c37a9314259d577aca91c6fd13765e5d6929037a3ef35586d0b0296: Status 404 returned error can't find the container with id 6ed438fb9c37a9314259d577aca91c6fd13765e5d6929037a3ef35586d0b0296 Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.449309 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerStarted","Data":"6ed438fb9c37a9314259d577aca91c6fd13765e5d6929037a3ef35586d0b0296"} Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.453523 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerStarted","Data":"ee285cde682fd3ee2d78c54cefcd8eb98634ecd91ce822877265e594545fe108"} Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.453562 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerStarted","Data":"2e9fa882b485b4cb5309fe0a2a2d2db25fed6080c8ce767b94dde09a4f33e1a9"} Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.891513 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:48:24 crc kubenswrapper[4888]: I1124 00:48:24.891850 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.469272 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerStarted","Data":"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102"} Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.469997 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.482649 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerStarted","Data":"ec323d623cb347bd09a2b25e20ca11b410d5f9d79a7685e048c188ef0c15c667"} Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.482701 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerStarted","Data":"8de4099978c784e388027c1d253ff368974310ef4fac664c5a8259064cd39a5a"} Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.486463 4888 generic.go:334] "Generic (PLEG): container finished" podID="de516985-13b9-49af-a46a-b072e62d6c25" containerID="1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155" exitCode=0 Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.486532 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerDied","Data":"1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155"} Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.498555 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.498709994 podStartE2EDuration="5.498542708s" podCreationTimestamp="2025-11-24 00:48:20 +0000 UTC" firstStartedPulling="2025-11-24 00:48:21.234742266 +0000 UTC m=+1403.817426310" lastFinishedPulling="2025-11-24 00:48:24.23457498 +0000 UTC m=+1406.817259024" observedRunningTime="2025-11-24 00:48:25.494904006 +0000 UTC m=+1408.077588050" watchObservedRunningTime="2025-11-24 00:48:25.498542708 +0000 UTC m=+1408.081226752" Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.903984 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.238:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:25 crc kubenswrapper[4888]: I1124 00:48:25.904229 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.238:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:26 crc kubenswrapper[4888]: I1124 00:48:26.505895 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerStarted","Data":"9f1d9abab17e134ac69708cd99eaa668c4bdd467af737ca28b2b5125b9b4bd32"} Nov 24 00:48:26 crc kubenswrapper[4888]: I1124 00:48:26.535476 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.352031965 podStartE2EDuration="4.535455087s" podCreationTimestamp="2025-11-24 00:48:22 +0000 UTC" firstStartedPulling="2025-11-24 00:48:23.589070767 +0000 UTC m=+1406.171754811" lastFinishedPulling="2025-11-24 00:48:25.772493889 +0000 UTC m=+1408.355177933" observedRunningTime="2025-11-24 00:48:26.531876577 +0000 UTC m=+1409.114560631" watchObservedRunningTime="2025-11-24 00:48:26.535455087 +0000 UTC m=+1409.118139131" Nov 24 00:48:27 crc kubenswrapper[4888]: I1124 00:48:27.516426 4888 generic.go:334] "Generic (PLEG): container finished" podID="cab6f695-53c2-43fe-bef0-fea107068d30" containerID="f03f887341c9e9500f0819e7e02b31aa263a5c14ba443321635f6bc69429f1d3" exitCode=0 Nov 24 00:48:27 crc kubenswrapper[4888]: I1124 00:48:27.516529 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmm2t" event={"ID":"cab6f695-53c2-43fe-bef0-fea107068d30","Type":"ContainerDied","Data":"f03f887341c9e9500f0819e7e02b31aa263a5c14ba443321635f6bc69429f1d3"} Nov 24 00:48:27 crc kubenswrapper[4888]: I1124 00:48:27.519164 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerStarted","Data":"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834"} Nov 24 00:48:28 crc kubenswrapper[4888]: I1124 00:48:28.963531 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.069422 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle\") pod \"cab6f695-53c2-43fe-bef0-fea107068d30\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.069545 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data\") pod \"cab6f695-53c2-43fe-bef0-fea107068d30\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.069586 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts\") pod \"cab6f695-53c2-43fe-bef0-fea107068d30\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.069632 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvftw\" (UniqueName: \"kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw\") pod \"cab6f695-53c2-43fe-bef0-fea107068d30\" (UID: \"cab6f695-53c2-43fe-bef0-fea107068d30\") " Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.078466 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw" (OuterVolumeSpecName: "kube-api-access-rvftw") pod "cab6f695-53c2-43fe-bef0-fea107068d30" (UID: "cab6f695-53c2-43fe-bef0-fea107068d30"). InnerVolumeSpecName "kube-api-access-rvftw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.085171 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts" (OuterVolumeSpecName: "scripts") pod "cab6f695-53c2-43fe-bef0-fea107068d30" (UID: "cab6f695-53c2-43fe-bef0-fea107068d30"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.109931 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cab6f695-53c2-43fe-bef0-fea107068d30" (UID: "cab6f695-53c2-43fe-bef0-fea107068d30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.129620 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data" (OuterVolumeSpecName: "config-data") pod "cab6f695-53c2-43fe-bef0-fea107068d30" (UID: "cab6f695-53c2-43fe-bef0-fea107068d30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.173103 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvftw\" (UniqueName: \"kubernetes.io/projected/cab6f695-53c2-43fe-bef0-fea107068d30-kube-api-access-rvftw\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.173144 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.173156 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.173167 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab6f695-53c2-43fe-bef0-fea107068d30-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.541476 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmm2t" event={"ID":"cab6f695-53c2-43fe-bef0-fea107068d30","Type":"ContainerDied","Data":"39fd31ecbc94fb1796ebc1d7a20aa75b2cf0e79199ad4da8ee3384ef3bc479d3"} Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.541531 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39fd31ecbc94fb1796ebc1d7a20aa75b2cf0e79199ad4da8ee3384ef3bc479d3" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.542068 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmm2t" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.544257 4888 generic.go:334] "Generic (PLEG): container finished" podID="de516985-13b9-49af-a46a-b072e62d6c25" containerID="0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834" exitCode=0 Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.544310 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerDied","Data":"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834"} Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.672884 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.673485 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.680875 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.750365 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.750853 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-log" containerID="cri-o://d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0" gracePeriod=30 Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.750899 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-api" containerID="cri-o://90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70" gracePeriod=30 Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.803143 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.816759 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:29 crc kubenswrapper[4888]: I1124 00:48:29.817020 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerName="nova-scheduler-scheduler" containerID="cri-o://1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" gracePeriod=30 Nov 24 00:48:30 crc kubenswrapper[4888]: I1124 00:48:30.560369 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerStarted","Data":"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf"} Nov 24 00:48:30 crc kubenswrapper[4888]: I1124 00:48:30.562672 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerID="d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0" exitCode=143 Nov 24 00:48:30 crc kubenswrapper[4888]: I1124 00:48:30.562904 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerDied","Data":"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0"} Nov 24 00:48:30 crc kubenswrapper[4888]: I1124 00:48:30.571099 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 00:48:30 crc kubenswrapper[4888]: I1124 00:48:30.657194 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hppc9" podStartSLOduration=3.258622225 podStartE2EDuration="7.657164906s" podCreationTimestamp="2025-11-24 00:48:23 +0000 UTC" firstStartedPulling="2025-11-24 00:48:25.498749244 +0000 UTC m=+1408.081433288" lastFinishedPulling="2025-11-24 00:48:29.897291925 +0000 UTC m=+1412.479975969" observedRunningTime="2025-11-24 00:48:30.582163289 +0000 UTC m=+1413.164847333" watchObservedRunningTime="2025-11-24 00:48:30.657164906 +0000 UTC m=+1413.239848970" Nov 24 00:48:31 crc kubenswrapper[4888]: I1124 00:48:31.570192 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" containerID="cri-o://bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f" gracePeriod=30 Nov 24 00:48:31 crc kubenswrapper[4888]: I1124 00:48:31.570310 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" containerID="cri-o://993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46" gracePeriod=30 Nov 24 00:48:32 crc kubenswrapper[4888]: I1124 00:48:32.585451 4888 generic.go:334] "Generic (PLEG): container finished" podID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerID="bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f" exitCode=143 Nov 24 00:48:32 crc kubenswrapper[4888]: I1124 00:48:32.585529 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerDied","Data":"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f"} Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.376945 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.458983 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.459078 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.459190 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r46jc\" (UniqueName: \"kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.459319 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.459407 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.459502 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle\") pod \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\" (UID: \"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3\") " Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.461139 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs" (OuterVolumeSpecName: "logs") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.464635 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc" (OuterVolumeSpecName: "kube-api-access-r46jc") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "kube-api-access-r46jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.489445 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.514672 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data" (OuterVolumeSpecName: "config-data") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.536284 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.537983 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" (UID: "fb4b0cb2-02f9-478c-be3d-126ea09bc9e3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.561957 4888 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.561993 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.562007 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.562017 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.562029 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.562040 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r46jc\" (UniqueName: \"kubernetes.io/projected/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3-kube-api-access-r46jc\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.597570 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerID="90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70" exitCode=0 Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.597617 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.597620 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerDied","Data":"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70"} Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.597731 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb4b0cb2-02f9-478c-be3d-126ea09bc9e3","Type":"ContainerDied","Data":"0b6fce4312ab7eceedcc5e8fc76eb34f05bae94d480e1814fffc80ad62f689f4"} Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.597747 4888 scope.go:117] "RemoveContainer" containerID="90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.692440 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.694506 4888 scope.go:117] "RemoveContainer" containerID="d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.701419 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.728121 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.728770 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab6f695-53c2-43fe-bef0-fea107068d30" containerName="nova-manage" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.728802 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab6f695-53c2-43fe-bef0-fea107068d30" containerName="nova-manage" Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.728863 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-api" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.728876 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-api" Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.728940 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-log" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.728952 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-log" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.729283 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-api" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.729315 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" containerName="nova-api-log" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.729336 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab6f695-53c2-43fe-bef0-fea107068d30" containerName="nova-manage" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.731087 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.733779 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.734330 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.734674 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.746044 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.756238 4888 scope.go:117] "RemoveContainer" containerID="90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70" Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.757557 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70\": container with ID starting with 90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70 not found: ID does not exist" containerID="90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.757604 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70"} err="failed to get container status \"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70\": rpc error: code = NotFound desc = could not find container \"90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70\": container with ID starting with 90dc97b13b74c7a9c3d633759e24029de15338ae51c10d19c40eb89b8d586d70 not found: ID does not exist" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.757638 4888 scope.go:117] "RemoveContainer" containerID="d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0" Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.759742 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0\": container with ID starting with d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0 not found: ID does not exist" containerID="d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.759786 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0"} err="failed to get container status \"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0\": rpc error: code = NotFound desc = could not find container \"d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0\": container with ID starting with d20fac041788f0d953ca9475fc34e0778bb8a80ad0e9039acc89a27f3b25f1e0 not found: ID does not exist" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.827432 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.827491 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868270 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-config-data\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868374 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868461 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868544 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg8fz\" (UniqueName: \"kubernetes.io/projected/f4060c08-c47b-483a-8a17-661a171187f2-kube-api-access-qg8fz\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868571 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4060c08-c47b-483a-8a17-661a171187f2-logs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.868612 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.932577 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 is running failed: container process not found" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.933073 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 is running failed: container process not found" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.933494 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 is running failed: container process not found" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 00:48:33 crc kubenswrapper[4888]: E1124 00:48:33.933533 4888 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerName="nova-scheduler-scheduler" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970268 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-config-data\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970328 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970425 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970489 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg8fz\" (UniqueName: \"kubernetes.io/projected/f4060c08-c47b-483a-8a17-661a171187f2-kube-api-access-qg8fz\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970533 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4060c08-c47b-483a-8a17-661a171187f2-logs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.970584 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.971036 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4060c08-c47b-483a-8a17-661a171187f2-logs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.975057 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-config-data\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.975321 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.976347 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.976669 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4060c08-c47b-483a-8a17-661a171187f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.977389 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:48:33 crc kubenswrapper[4888]: I1124 00:48:33.987212 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg8fz\" (UniqueName: \"kubernetes.io/projected/f4060c08-c47b-483a-8a17-661a171187f2-kube-api-access-qg8fz\") pod \"nova-api-0\" (UID: \"f4060c08-c47b-483a-8a17-661a171187f2\") " pod="openstack/nova-api-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.055231 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.071635 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle\") pod \"10f840a2-3864-489a-a694-ac89fa5dadaa\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.072011 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8cwv\" (UniqueName: \"kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv\") pod \"10f840a2-3864-489a-a694-ac89fa5dadaa\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.072044 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data\") pod \"10f840a2-3864-489a-a694-ac89fa5dadaa\" (UID: \"10f840a2-3864-489a-a694-ac89fa5dadaa\") " Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.077062 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv" (OuterVolumeSpecName: "kube-api-access-h8cwv") pod "10f840a2-3864-489a-a694-ac89fa5dadaa" (UID: "10f840a2-3864-489a-a694-ac89fa5dadaa"). InnerVolumeSpecName "kube-api-access-h8cwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.116424 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10f840a2-3864-489a-a694-ac89fa5dadaa" (UID: "10f840a2-3864-489a-a694-ac89fa5dadaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.127848 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data" (OuterVolumeSpecName: "config-data") pod "10f840a2-3864-489a-a694-ac89fa5dadaa" (UID: "10f840a2-3864-489a-a694-ac89fa5dadaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.174720 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8cwv\" (UniqueName: \"kubernetes.io/projected/10f840a2-3864-489a-a694-ac89fa5dadaa-kube-api-access-h8cwv\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.174818 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.174832 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f840a2-3864-489a-a694-ac89fa5dadaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.270734 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb4b0cb2-02f9-478c-be3d-126ea09bc9e3" path="/var/lib/kubelet/pods/fb4b0cb2-02f9-478c-be3d-126ea09bc9e3/volumes" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.569682 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 00:48:34 crc kubenswrapper[4888]: W1124 00:48:34.576248 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4060c08_c47b_483a_8a17_661a171187f2.slice/crio-03d1bc289d07b7bcd2568ed04a9d49c835ff28c66ef677debcb39c0a9ba830ae WatchSource:0}: Error finding container 03d1bc289d07b7bcd2568ed04a9d49c835ff28c66ef677debcb39c0a9ba830ae: Status 404 returned error can't find the container with id 03d1bc289d07b7bcd2568ed04a9d49c835ff28c66ef677debcb39c0a9ba830ae Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.624237 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f4060c08-c47b-483a-8a17-661a171187f2","Type":"ContainerStarted","Data":"03d1bc289d07b7bcd2568ed04a9d49c835ff28c66ef677debcb39c0a9ba830ae"} Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.634418 4888 generic.go:334] "Generic (PLEG): container finished" podID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" exitCode=0 Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.634470 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"10f840a2-3864-489a-a694-ac89fa5dadaa","Type":"ContainerDied","Data":"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8"} Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.634497 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"10f840a2-3864-489a-a694-ac89fa5dadaa","Type":"ContainerDied","Data":"213f64ea9be191685a39fa1358268d98756b019c29117c8018aa92c3a954d7b3"} Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.634523 4888 scope.go:117] "RemoveContainer" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.634753 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.692558 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.700396 4888 scope.go:117] "RemoveContainer" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" Nov 24 00:48:34 crc kubenswrapper[4888]: E1124 00:48:34.701262 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8\": container with ID starting with 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 not found: ID does not exist" containerID="1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.701323 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8"} err="failed to get container status \"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8\": rpc error: code = NotFound desc = could not find container \"1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8\": container with ID starting with 1d7dcaf3bd3fca350d46c4eb821c247090e0b14503203a73621259590b117ee8 not found: ID does not exist" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.707346 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.719278 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:34 crc kubenswrapper[4888]: E1124 00:48:34.719672 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerName="nova-scheduler-scheduler" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.719690 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerName="nova-scheduler-scheduler" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.719910 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" containerName="nova-scheduler-scheduler" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.720612 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.729614 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.734106 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.747253 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.235:8775/\": read tcp 10.217.0.2:39208->10.217.0.235:8775: read: connection reset by peer" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.747311 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.235:8775/\": read tcp 10.217.0.2:39206->10.217.0.235:8775: read: connection reset by peer" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.786549 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7lpc\" (UniqueName: \"kubernetes.io/projected/894a3216-d75f-4229-849f-4b7f4e942d96-kube-api-access-s7lpc\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.786607 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-config-data\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.786684 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.888089 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hppc9" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="registry-server" probeResult="failure" output=< Nov 24 00:48:34 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:48:34 crc kubenswrapper[4888]: > Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.888277 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7lpc\" (UniqueName: \"kubernetes.io/projected/894a3216-d75f-4229-849f-4b7f4e942d96-kube-api-access-s7lpc\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.888672 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-config-data\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.888793 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.895375 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-config-data\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.896073 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894a3216-d75f-4229-849f-4b7f4e942d96-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:34 crc kubenswrapper[4888]: I1124 00:48:34.905778 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7lpc\" (UniqueName: \"kubernetes.io/projected/894a3216-d75f-4229-849f-4b7f4e942d96-kube-api-access-s7lpc\") pod \"nova-scheduler-0\" (UID: \"894a3216-d75f-4229-849f-4b7f4e942d96\") " pod="openstack/nova-scheduler-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.052688 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.215385 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.299446 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle\") pod \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.299491 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs\") pod \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.299544 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data\") pod \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.299567 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs\") pod \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.299843 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm75s\" (UniqueName: \"kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s\") pod \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\" (UID: \"12a63eaf-6ecf-4c8c-b1a7-228114ae582a\") " Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.300473 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs" (OuterVolumeSpecName: "logs") pod "12a63eaf-6ecf-4c8c-b1a7-228114ae582a" (UID: "12a63eaf-6ecf-4c8c-b1a7-228114ae582a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.310039 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s" (OuterVolumeSpecName: "kube-api-access-dm75s") pod "12a63eaf-6ecf-4c8c-b1a7-228114ae582a" (UID: "12a63eaf-6ecf-4c8c-b1a7-228114ae582a"). InnerVolumeSpecName "kube-api-access-dm75s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.343233 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12a63eaf-6ecf-4c8c-b1a7-228114ae582a" (UID: "12a63eaf-6ecf-4c8c-b1a7-228114ae582a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.357459 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data" (OuterVolumeSpecName: "config-data") pod "12a63eaf-6ecf-4c8c-b1a7-228114ae582a" (UID: "12a63eaf-6ecf-4c8c-b1a7-228114ae582a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.378806 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "12a63eaf-6ecf-4c8c-b1a7-228114ae582a" (UID: "12a63eaf-6ecf-4c8c-b1a7-228114ae582a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.402956 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.402991 4888 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.403005 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.403015 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.403026 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm75s\" (UniqueName: \"kubernetes.io/projected/12a63eaf-6ecf-4c8c-b1a7-228114ae582a-kube-api-access-dm75s\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:35 crc kubenswrapper[4888]: W1124 00:48:35.509376 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod894a3216_d75f_4229_849f_4b7f4e942d96.slice/crio-d12bc45df3081c4ac8d14d9f3e8114b362a5ecd96d669b6125be87e1327f0fe8 WatchSource:0}: Error finding container d12bc45df3081c4ac8d14d9f3e8114b362a5ecd96d669b6125be87e1327f0fe8: Status 404 returned error can't find the container with id d12bc45df3081c4ac8d14d9f3e8114b362a5ecd96d669b6125be87e1327f0fe8 Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.510790 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.648998 4888 generic.go:334] "Generic (PLEG): container finished" podID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerID="993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46" exitCode=0 Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.649069 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.649110 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerDied","Data":"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46"} Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.649386 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12a63eaf-6ecf-4c8c-b1a7-228114ae582a","Type":"ContainerDied","Data":"cc71647271ff883ba70eef8ea7af797bf9ee053198de894e06267a80b8ebc655"} Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.649405 4888 scope.go:117] "RemoveContainer" containerID="993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.660005 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f4060c08-c47b-483a-8a17-661a171187f2","Type":"ContainerStarted","Data":"c2285be042f82be8de39287144b2ac9c5356823d31a3e3d2d863cd73642b5a38"} Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.660426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f4060c08-c47b-483a-8a17-661a171187f2","Type":"ContainerStarted","Data":"03905d34dbf37a90421c90233982507038172ade90df06643ef06f578578556b"} Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.669594 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"894a3216-d75f-4229-849f-4b7f4e942d96","Type":"ContainerStarted","Data":"d12bc45df3081c4ac8d14d9f3e8114b362a5ecd96d669b6125be87e1327f0fe8"} Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.696995 4888 scope.go:117] "RemoveContainer" containerID="bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.698876 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.698857435 podStartE2EDuration="2.698857435s" podCreationTimestamp="2025-11-24 00:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:35.689408341 +0000 UTC m=+1418.272092385" watchObservedRunningTime="2025-11-24 00:48:35.698857435 +0000 UTC m=+1418.281541469" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.720339 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.736518 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.743370 4888 scope.go:117] "RemoveContainer" containerID="993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46" Nov 24 00:48:35 crc kubenswrapper[4888]: E1124 00:48:35.743766 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46\": container with ID starting with 993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46 not found: ID does not exist" containerID="993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.743793 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46"} err="failed to get container status \"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46\": rpc error: code = NotFound desc = could not find container \"993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46\": container with ID starting with 993fa1461213bbdcb6c4e7781897f5ea58e1949f4deed5c6832722f6b5ba0d46 not found: ID does not exist" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.743830 4888 scope.go:117] "RemoveContainer" containerID="bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f" Nov 24 00:48:35 crc kubenswrapper[4888]: E1124 00:48:35.744024 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f\": container with ID starting with bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f not found: ID does not exist" containerID="bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.744047 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f"} err="failed to get container status \"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f\": rpc error: code = NotFound desc = could not find container \"bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f\": container with ID starting with bd7646cf976fcc555e9c880b202ae9074338b2a96afda30ead3df25d97f4826f not found: ID does not exist" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.769302 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:35 crc kubenswrapper[4888]: E1124 00:48:35.769877 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.769899 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" Nov 24 00:48:35 crc kubenswrapper[4888]: E1124 00:48:35.769943 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.769952 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.770190 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-metadata" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.770219 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" containerName="nova-metadata-log" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.771656 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.774084 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.774404 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.792845 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:35 crc kubenswrapper[4888]: E1124 00:48:35.867784 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12a63eaf_6ecf_4c8c_b1a7_228114ae582a.slice/crio-cc71647271ff883ba70eef8ea7af797bf9ee053198de894e06267a80b8ebc655\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12a63eaf_6ecf_4c8c_b1a7_228114ae582a.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.912147 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.912186 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.912218 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxxbw\" (UniqueName: \"kubernetes.io/projected/bfb6aad7-f065-42f1-9021-f5debe134c7d-kube-api-access-pxxbw\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.912235 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-config-data\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:35 crc kubenswrapper[4888]: I1124 00:48:35.912252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfb6aad7-f065-42f1-9021-f5debe134c7d-logs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.014533 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.014574 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.014603 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-config-data\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.014648 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxxbw\" (UniqueName: \"kubernetes.io/projected/bfb6aad7-f065-42f1-9021-f5debe134c7d-kube-api-access-pxxbw\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.014670 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfb6aad7-f065-42f1-9021-f5debe134c7d-logs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.015135 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfb6aad7-f065-42f1-9021-f5debe134c7d-logs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.019895 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-config-data\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.020289 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.020411 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb6aad7-f065-42f1-9021-f5debe134c7d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.030587 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxxbw\" (UniqueName: \"kubernetes.io/projected/bfb6aad7-f065-42f1-9021-f5debe134c7d-kube-api-access-pxxbw\") pod \"nova-metadata-0\" (UID: \"bfb6aad7-f065-42f1-9021-f5debe134c7d\") " pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.092944 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.264849 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f840a2-3864-489a-a694-ac89fa5dadaa" path="/var/lib/kubelet/pods/10f840a2-3864-489a-a694-ac89fa5dadaa/volumes" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.266034 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12a63eaf-6ecf-4c8c-b1a7-228114ae582a" path="/var/lib/kubelet/pods/12a63eaf-6ecf-4c8c-b1a7-228114ae582a/volumes" Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.566731 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.695973 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"894a3216-d75f-4229-849f-4b7f4e942d96","Type":"ContainerStarted","Data":"77da96c3ed38e7d7ffd9122ac2fd0b2eed761d935a9c1b650e7c2794f55a702e"} Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.704600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfb6aad7-f065-42f1-9021-f5debe134c7d","Type":"ContainerStarted","Data":"37cfe15c1f8e890b5b4c77b87748af80bce88fdc97c5b3e23847c1b0823dcaf8"} Nov 24 00:48:36 crc kubenswrapper[4888]: I1124 00:48:36.731498 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.731475043 podStartE2EDuration="2.731475043s" podCreationTimestamp="2025-11-24 00:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:36.71780228 +0000 UTC m=+1419.300486344" watchObservedRunningTime="2025-11-24 00:48:36.731475043 +0000 UTC m=+1419.314159107" Nov 24 00:48:37 crc kubenswrapper[4888]: I1124 00:48:37.714519 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfb6aad7-f065-42f1-9021-f5debe134c7d","Type":"ContainerStarted","Data":"d8c26713ceec30173898f3556fe7d7633c9e9849fc799bbb991c79e656dd4d05"} Nov 24 00:48:37 crc kubenswrapper[4888]: I1124 00:48:37.714968 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfb6aad7-f065-42f1-9021-f5debe134c7d","Type":"ContainerStarted","Data":"a04a142b0a4d1d344c0fefddc502a18eb5f44fc4424bbd8ca92e4b604b3603b9"} Nov 24 00:48:37 crc kubenswrapper[4888]: I1124 00:48:37.743061 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.743043513 podStartE2EDuration="2.743043513s" podCreationTimestamp="2025-11-24 00:48:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:48:37.733737043 +0000 UTC m=+1420.316421087" watchObservedRunningTime="2025-11-24 00:48:37.743043513 +0000 UTC m=+1420.325727557" Nov 24 00:48:40 crc kubenswrapper[4888]: I1124 00:48:40.053294 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 00:48:41 crc kubenswrapper[4888]: I1124 00:48:41.093099 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:48:41 crc kubenswrapper[4888]: I1124 00:48:41.093422 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 00:48:43 crc kubenswrapper[4888]: I1124 00:48:43.886547 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:43 crc kubenswrapper[4888]: I1124 00:48:43.956318 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:44 crc kubenswrapper[4888]: I1124 00:48:44.056033 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:48:44 crc kubenswrapper[4888]: I1124 00:48:44.056129 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 00:48:44 crc kubenswrapper[4888]: I1124 00:48:44.141636 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.054006 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.068920 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f4060c08-c47b-483a-8a17-661a171187f2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.243:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.069017 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f4060c08-c47b-483a-8a17-661a171187f2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.243:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.090020 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.819445 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hppc9" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="registry-server" containerID="cri-o://61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf" gracePeriod=2 Nov 24 00:48:45 crc kubenswrapper[4888]: I1124 00:48:45.859589 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.093378 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.093738 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.405799 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.467532 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content\") pod \"de516985-13b9-49af-a46a-b072e62d6c25\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.467650 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities\") pod \"de516985-13b9-49af-a46a-b072e62d6c25\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.467746 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dltjs\" (UniqueName: \"kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs\") pod \"de516985-13b9-49af-a46a-b072e62d6c25\" (UID: \"de516985-13b9-49af-a46a-b072e62d6c25\") " Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.468718 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities" (OuterVolumeSpecName: "utilities") pod "de516985-13b9-49af-a46a-b072e62d6c25" (UID: "de516985-13b9-49af-a46a-b072e62d6c25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.496430 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs" (OuterVolumeSpecName: "kube-api-access-dltjs") pod "de516985-13b9-49af-a46a-b072e62d6c25" (UID: "de516985-13b9-49af-a46a-b072e62d6c25"). InnerVolumeSpecName "kube-api-access-dltjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.569838 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.569873 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dltjs\" (UniqueName: \"kubernetes.io/projected/de516985-13b9-49af-a46a-b072e62d6c25-kube-api-access-dltjs\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.595113 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de516985-13b9-49af-a46a-b072e62d6c25" (UID: "de516985-13b9-49af-a46a-b072e62d6c25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.672501 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de516985-13b9-49af-a46a-b072e62d6c25-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.847652 4888 generic.go:334] "Generic (PLEG): container finished" podID="de516985-13b9-49af-a46a-b072e62d6c25" containerID="61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf" exitCode=0 Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.847700 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hppc9" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.847754 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerDied","Data":"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf"} Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.847844 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hppc9" event={"ID":"de516985-13b9-49af-a46a-b072e62d6c25","Type":"ContainerDied","Data":"6ed438fb9c37a9314259d577aca91c6fd13765e5d6929037a3ef35586d0b0296"} Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.847868 4888 scope.go:117] "RemoveContainer" containerID="61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.880905 4888 scope.go:117] "RemoveContainer" containerID="0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.903837 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.922091 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hppc9"] Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.927927 4888 scope.go:117] "RemoveContainer" containerID="1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.976455 4888 scope.go:117] "RemoveContainer" containerID="61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf" Nov 24 00:48:46 crc kubenswrapper[4888]: E1124 00:48:46.976928 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf\": container with ID starting with 61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf not found: ID does not exist" containerID="61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.976977 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf"} err="failed to get container status \"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf\": rpc error: code = NotFound desc = could not find container \"61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf\": container with ID starting with 61f3c5a3c933cceb0c01601bad57833d6182703f0ab3c2aa029a9619101b3dbf not found: ID does not exist" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.977009 4888 scope.go:117] "RemoveContainer" containerID="0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834" Nov 24 00:48:46 crc kubenswrapper[4888]: E1124 00:48:46.977346 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834\": container with ID starting with 0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834 not found: ID does not exist" containerID="0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.977378 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834"} err="failed to get container status \"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834\": rpc error: code = NotFound desc = could not find container \"0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834\": container with ID starting with 0e59b229b956e58a4caf3914d87d56a865fc92bdea7f6f0106a9ce10e2c8d834 not found: ID does not exist" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.977400 4888 scope.go:117] "RemoveContainer" containerID="1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155" Nov 24 00:48:46 crc kubenswrapper[4888]: E1124 00:48:46.977673 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155\": container with ID starting with 1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155 not found: ID does not exist" containerID="1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155" Nov 24 00:48:46 crc kubenswrapper[4888]: I1124 00:48:46.977704 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155"} err="failed to get container status \"1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155\": rpc error: code = NotFound desc = could not find container \"1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155\": container with ID starting with 1e6a9116b50187ec915c5bbcc8d3d414f22bfbfb531848872b5e63644a3c5155 not found: ID does not exist" Nov 24 00:48:47 crc kubenswrapper[4888]: I1124 00:48:47.105981 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfb6aad7-f065-42f1-9021-f5debe134c7d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:47 crc kubenswrapper[4888]: I1124 00:48:47.106027 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfb6aad7-f065-42f1-9021-f5debe134c7d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 00:48:48 crc kubenswrapper[4888]: I1124 00:48:48.266451 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de516985-13b9-49af-a46a-b072e62d6c25" path="/var/lib/kubelet/pods/de516985-13b9-49af-a46a-b072e62d6c25/volumes" Nov 24 00:48:50 crc kubenswrapper[4888]: I1124 00:48:50.753521 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 00:48:53 crc kubenswrapper[4888]: I1124 00:48:53.591092 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:48:53 crc kubenswrapper[4888]: I1124 00:48:53.591416 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.067408 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.068075 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.069312 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.072917 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.971860 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 00:48:54 crc kubenswrapper[4888]: I1124 00:48:54.980564 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.112269 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.112446 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4678859e-43e2-4d0b-ae68-323ec7239ad7" containerName="kube-state-metrics" containerID="cri-o://1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988" gracePeriod=30 Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.267554 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.269367 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="1565ef5a-8292-4a7e-9484-c6483c547353" containerName="mysqld-exporter" containerID="cri-o://4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5" gracePeriod=30 Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.665721 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.762050 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.785403 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28kxf\" (UniqueName: \"kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf\") pod \"4678859e-43e2-4d0b-ae68-323ec7239ad7\" (UID: \"4678859e-43e2-4d0b-ae68-323ec7239ad7\") " Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.793091 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf" (OuterVolumeSpecName: "kube-api-access-28kxf") pod "4678859e-43e2-4d0b-ae68-323ec7239ad7" (UID: "4678859e-43e2-4d0b-ae68-323ec7239ad7"). InnerVolumeSpecName "kube-api-access-28kxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.887682 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzw7v\" (UniqueName: \"kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v\") pod \"1565ef5a-8292-4a7e-9484-c6483c547353\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.887979 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle\") pod \"1565ef5a-8292-4a7e-9484-c6483c547353\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.888022 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data\") pod \"1565ef5a-8292-4a7e-9484-c6483c547353\" (UID: \"1565ef5a-8292-4a7e-9484-c6483c547353\") " Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.888390 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28kxf\" (UniqueName: \"kubernetes.io/projected/4678859e-43e2-4d0b-ae68-323ec7239ad7-kube-api-access-28kxf\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.892220 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v" (OuterVolumeSpecName: "kube-api-access-vzw7v") pod "1565ef5a-8292-4a7e-9484-c6483c547353" (UID: "1565ef5a-8292-4a7e-9484-c6483c547353"). InnerVolumeSpecName "kube-api-access-vzw7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.912982 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1565ef5a-8292-4a7e-9484-c6483c547353" (UID: "1565ef5a-8292-4a7e-9484-c6483c547353"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.950668 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data" (OuterVolumeSpecName: "config-data") pod "1565ef5a-8292-4a7e-9484-c6483c547353" (UID: "1565ef5a-8292-4a7e-9484-c6483c547353"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.983982 4888 generic.go:334] "Generic (PLEG): container finished" podID="4678859e-43e2-4d0b-ae68-323ec7239ad7" containerID="1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988" exitCode=2 Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.984050 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4678859e-43e2-4d0b-ae68-323ec7239ad7","Type":"ContainerDied","Data":"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988"} Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.984084 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4678859e-43e2-4d0b-ae68-323ec7239ad7","Type":"ContainerDied","Data":"7c3fb792900cd2feb6472fefae4de29b62fc2a5697b0f5b9476c761a0c966088"} Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.984095 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.984101 4888 scope.go:117] "RemoveContainer" containerID="1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.986483 4888 generic.go:334] "Generic (PLEG): container finished" podID="1565ef5a-8292-4a7e-9484-c6483c547353" containerID="4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5" exitCode=2 Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.987511 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.996937 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"1565ef5a-8292-4a7e-9484-c6483c547353","Type":"ContainerDied","Data":"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5"} Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.997019 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"1565ef5a-8292-4a7e-9484-c6483c547353","Type":"ContainerDied","Data":"c0636154a0b16eb63220790ed87311000f676387b8cf2ba48b5d397820dc0c8d"} Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.999309 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzw7v\" (UniqueName: \"kubernetes.io/projected/1565ef5a-8292-4a7e-9484-c6483c547353-kube-api-access-vzw7v\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.999406 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:55 crc kubenswrapper[4888]: I1124 00:48:55.999460 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1565ef5a-8292-4a7e-9484-c6483c547353-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.014639 4888 scope.go:117] "RemoveContainer" containerID="1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988" Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.015078 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988\": container with ID starting with 1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988 not found: ID does not exist" containerID="1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.015114 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988"} err="failed to get container status \"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988\": rpc error: code = NotFound desc = could not find container \"1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988\": container with ID starting with 1f9051bd872f4cbb262b562a5b7536240fe0015925d735cef441e3b5b3208988 not found: ID does not exist" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.015136 4888 scope.go:117] "RemoveContainer" containerID="4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.048553 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.076354 4888 scope.go:117] "RemoveContainer" containerID="4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.084965 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.085369 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5\": container with ID starting with 4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5 not found: ID does not exist" containerID="4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.085404 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5"} err="failed to get container status \"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5\": rpc error: code = NotFound desc = could not find container \"4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5\": container with ID starting with 4857b2edd1397753c3babad69fa88eb48e82ccaafdffc02c965298f1e33439b5 not found: ID does not exist" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.096937 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.106248 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109135 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.109649 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1565ef5a-8292-4a7e-9484-c6483c547353" containerName="mysqld-exporter" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109672 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1565ef5a-8292-4a7e-9484-c6483c547353" containerName="mysqld-exporter" Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.109695 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="extract-utilities" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109705 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="extract-utilities" Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.109723 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4678859e-43e2-4d0b-ae68-323ec7239ad7" containerName="kube-state-metrics" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109731 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4678859e-43e2-4d0b-ae68-323ec7239ad7" containerName="kube-state-metrics" Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.109784 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="registry-server" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109793 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="registry-server" Nov 24 00:48:56 crc kubenswrapper[4888]: E1124 00:48:56.109842 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="extract-content" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.109851 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="extract-content" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.110133 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="de516985-13b9-49af-a46a-b072e62d6c25" containerName="registry-server" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.110166 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4678859e-43e2-4d0b-ae68-323ec7239ad7" containerName="kube-state-metrics" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.110187 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1565ef5a-8292-4a7e-9484-c6483c547353" containerName="mysqld-exporter" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.111250 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.111334 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.113279 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.113451 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.117390 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.119869 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.131534 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.150920 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.153591 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.161279 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.161541 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.163941 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.202497 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-config-data\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.202588 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.202639 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg2pg\" (UniqueName: \"kubernetes.io/projected/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-kube-api-access-zg2pg\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.202839 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.274504 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1565ef5a-8292-4a7e-9484-c6483c547353" path="/var/lib/kubelet/pods/1565ef5a-8292-4a7e-9484-c6483c547353/volumes" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.275198 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4678859e-43e2-4d0b-ae68-323ec7239ad7" path="/var/lib/kubelet/pods/4678859e-43e2-4d0b-ae68-323ec7239ad7/volumes" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307138 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-config-data\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307195 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307224 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg2pg\" (UniqueName: \"kubernetes.io/projected/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-kube-api-access-zg2pg\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307316 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307383 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307432 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307478 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cllm\" (UniqueName: \"kubernetes.io/projected/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-api-access-8cllm\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.307497 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.314018 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.317676 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.322119 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-config-data\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.330157 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg2pg\" (UniqueName: \"kubernetes.io/projected/f25aa6c8-4f87-4d83-8563-fc22a4bfa695-kube-api-access-zg2pg\") pod \"mysqld-exporter-0\" (UID: \"f25aa6c8-4f87-4d83-8563-fc22a4bfa695\") " pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.409218 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.409297 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cllm\" (UniqueName: \"kubernetes.io/projected/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-api-access-8cllm\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.409319 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.409420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.413788 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.413873 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.414536 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.431541 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cllm\" (UniqueName: \"kubernetes.io/projected/fe0c6981-4d50-4dba-b7a4-cabf7beb1947-kube-api-access-8cllm\") pod \"kube-state-metrics-0\" (UID: \"fe0c6981-4d50-4dba-b7a4-cabf7beb1947\") " pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.433794 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.475187 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.918881 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 24 00:48:56 crc kubenswrapper[4888]: W1124 00:48:56.981493 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe0c6981_4d50_4dba_b7a4_cabf7beb1947.slice/crio-d618149e64a77378fa73d374822f941c1c747df58d3e0ff20415d73740750fd6 WatchSource:0}: Error finding container d618149e64a77378fa73d374822f941c1c747df58d3e0ff20415d73740750fd6: Status 404 returned error can't find the container with id d618149e64a77378fa73d374822f941c1c747df58d3e0ff20415d73740750fd6 Nov 24 00:48:56 crc kubenswrapper[4888]: I1124 00:48:56.991178 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.002079 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f25aa6c8-4f87-4d83-8563-fc22a4bfa695","Type":"ContainerStarted","Data":"44e37f06395040fce3cc5513d8e5d77f60a4c51e34c8559e4d6b6ea1318daa7f"} Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.004018 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe0c6981-4d50-4dba-b7a4-cabf7beb1947","Type":"ContainerStarted","Data":"d618149e64a77378fa73d374822f941c1c747df58d3e0ff20415d73740750fd6"} Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.015639 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.140793 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.141412 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-central-agent" containerID="cri-o://6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99" gracePeriod=30 Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.141478 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="proxy-httpd" containerID="cri-o://f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102" gracePeriod=30 Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.141490 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="sg-core" containerID="cri-o://ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb" gracePeriod=30 Nov 24 00:48:57 crc kubenswrapper[4888]: I1124 00:48:57.141508 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-notification-agent" containerID="cri-o://6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0" gracePeriod=30 Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.023897 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f25aa6c8-4f87-4d83-8563-fc22a4bfa695","Type":"ContainerStarted","Data":"af11058b9b6abacbade102c96d57453a65dd77e7c9da8512b5e3f0204fce92d6"} Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.027347 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe0c6981-4d50-4dba-b7a4-cabf7beb1947","Type":"ContainerStarted","Data":"c8fcc823a7740dc1b768e2e6d912fe43629f899ef4717e85733ec906a19e1ef2"} Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.027452 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030024 4888 generic.go:334] "Generic (PLEG): container finished" podID="013921b1-bc4d-4433-95de-8603403ab9e2" containerID="f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102" exitCode=0 Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030126 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerDied","Data":"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102"} Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030183 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerDied","Data":"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb"} Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030143 4888 generic.go:334] "Generic (PLEG): container finished" podID="013921b1-bc4d-4433-95de-8603403ab9e2" containerID="ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb" exitCode=2 Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030215 4888 generic.go:334] "Generic (PLEG): container finished" podID="013921b1-bc4d-4433-95de-8603403ab9e2" containerID="6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99" exitCode=0 Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.030319 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerDied","Data":"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99"} Nov 24 00:48:58 crc kubenswrapper[4888]: I1124 00:48:58.058743 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=1.43590641 podStartE2EDuration="2.058720337s" podCreationTimestamp="2025-11-24 00:48:56 +0000 UTC" firstStartedPulling="2025-11-24 00:48:56.934170948 +0000 UTC m=+1439.516854992" lastFinishedPulling="2025-11-24 00:48:57.556984875 +0000 UTC m=+1440.139668919" observedRunningTime="2025-11-24 00:48:58.045367154 +0000 UTC m=+1440.628051208" watchObservedRunningTime="2025-11-24 00:48:58.058720337 +0000 UTC m=+1440.641404411" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.593856 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.625874 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.206392585 podStartE2EDuration="3.625860645s" podCreationTimestamp="2025-11-24 00:48:56 +0000 UTC" firstStartedPulling="2025-11-24 00:48:56.985046571 +0000 UTC m=+1439.567730625" lastFinishedPulling="2025-11-24 00:48:57.404514641 +0000 UTC m=+1439.987198685" observedRunningTime="2025-11-24 00:48:58.074651633 +0000 UTC m=+1440.657335677" watchObservedRunningTime="2025-11-24 00:48:59.625860645 +0000 UTC m=+1442.208544689" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681060 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681106 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681178 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681196 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681219 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dwwg\" (UniqueName: \"kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681277 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.681310 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml\") pod \"013921b1-bc4d-4433-95de-8603403ab9e2\" (UID: \"013921b1-bc4d-4433-95de-8603403ab9e2\") " Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.682392 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.682645 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.688049 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts" (OuterVolumeSpecName: "scripts") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.688445 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg" (OuterVolumeSpecName: "kube-api-access-8dwwg") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "kube-api-access-8dwwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.711087 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.773377 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784181 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784328 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784384 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784436 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/013921b1-bc4d-4433-95de-8603403ab9e2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784486 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.784537 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dwwg\" (UniqueName: \"kubernetes.io/projected/013921b1-bc4d-4433-95de-8603403ab9e2-kube-api-access-8dwwg\") on node \"crc\" DevicePath \"\"" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.813685 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data" (OuterVolumeSpecName: "config-data") pod "013921b1-bc4d-4433-95de-8603403ab9e2" (UID: "013921b1-bc4d-4433-95de-8603403ab9e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:48:59 crc kubenswrapper[4888]: I1124 00:48:59.889396 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013921b1-bc4d-4433-95de-8603403ab9e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.056861 4888 generic.go:334] "Generic (PLEG): container finished" podID="013921b1-bc4d-4433-95de-8603403ab9e2" containerID="6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0" exitCode=0 Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.056919 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerDied","Data":"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0"} Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.056958 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"013921b1-bc4d-4433-95de-8603403ab9e2","Type":"ContainerDied","Data":"66098b326642ee723ad9853f55baf3d2bd09cddc0663303c5760e47513fe7bd7"} Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.056990 4888 scope.go:117] "RemoveContainer" containerID="f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.057201 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.081495 4888 scope.go:117] "RemoveContainer" containerID="ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.113632 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.122505 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146294 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.146688 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="sg-core" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146703 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="sg-core" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.146726 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="proxy-httpd" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146732 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="proxy-httpd" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.146742 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-central-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146748 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-central-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.146776 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-notification-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146782 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-notification-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146984 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-notification-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.146997 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="ceilometer-central-agent" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.147009 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="proxy-httpd" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.147019 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" containerName="sg-core" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.148661 4888 scope.go:117] "RemoveContainer" containerID="6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.148803 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.151262 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.151510 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.151692 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.163645 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.184083 4888 scope.go:117] "RemoveContainer" containerID="6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.205857 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.205916 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.205974 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.206014 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.206037 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.206092 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.206162 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq827\" (UniqueName: \"kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.206235 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.207681 4888 scope.go:117] "RemoveContainer" containerID="f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.208557 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102\": container with ID starting with f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102 not found: ID does not exist" containerID="f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.208599 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102"} err="failed to get container status \"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102\": rpc error: code = NotFound desc = could not find container \"f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102\": container with ID starting with f6a0747db554b0789f67fae0ee4b668df0e2c6b9d14f5469dd3a305f13adb102 not found: ID does not exist" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.208630 4888 scope.go:117] "RemoveContainer" containerID="ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.209335 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb\": container with ID starting with ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb not found: ID does not exist" containerID="ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.209372 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb"} err="failed to get container status \"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb\": rpc error: code = NotFound desc = could not find container \"ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb\": container with ID starting with ac8fc5ae31c227fc8af8ca994e51f354aba5437543fa7cc925aa729158d335bb not found: ID does not exist" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.209391 4888 scope.go:117] "RemoveContainer" containerID="6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.209828 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0\": container with ID starting with 6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0 not found: ID does not exist" containerID="6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.209864 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0"} err="failed to get container status \"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0\": rpc error: code = NotFound desc = could not find container \"6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0\": container with ID starting with 6fb9ec619bd5a9d33631b315d5fa2ce161617771d6e83edd014b9ad75a86e6e0 not found: ID does not exist" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.209889 4888 scope.go:117] "RemoveContainer" containerID="6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99" Nov 24 00:49:00 crc kubenswrapper[4888]: E1124 00:49:00.210388 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99\": container with ID starting with 6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99 not found: ID does not exist" containerID="6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.210435 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99"} err="failed to get container status \"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99\": rpc error: code = NotFound desc = could not find container \"6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99\": container with ID starting with 6ffe1abe20d956934ca732235bb3d164282443e24011ccd1381a3654c8539d99 not found: ID does not exist" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.263710 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="013921b1-bc4d-4433-95de-8603403ab9e2" path="/var/lib/kubelet/pods/013921b1-bc4d-4433-95de-8603403ab9e2/volumes" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308000 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308052 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308099 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308136 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308157 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308185 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308224 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq827\" (UniqueName: \"kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.308313 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.309543 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.309885 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.312634 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.312724 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.312996 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.313845 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.314725 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.327640 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq827\" (UniqueName: \"kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827\") pod \"ceilometer-0\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " pod="openstack/ceilometer-0" Nov 24 00:49:00 crc kubenswrapper[4888]: I1124 00:49:00.472501 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:01 crc kubenswrapper[4888]: I1124 00:49:01.001364 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:01 crc kubenswrapper[4888]: W1124 00:49:01.009273 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a956701_1723_4960_8b2f_cd581eeb4855.slice/crio-ccfd4e883d77908cdec2d69373f9d0b8c5f4a36e90b83266ffb88105e989131b WatchSource:0}: Error finding container ccfd4e883d77908cdec2d69373f9d0b8c5f4a36e90b83266ffb88105e989131b: Status 404 returned error can't find the container with id ccfd4e883d77908cdec2d69373f9d0b8c5f4a36e90b83266ffb88105e989131b Nov 24 00:49:01 crc kubenswrapper[4888]: I1124 00:49:01.072050 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerStarted","Data":"ccfd4e883d77908cdec2d69373f9d0b8c5f4a36e90b83266ffb88105e989131b"} Nov 24 00:49:02 crc kubenswrapper[4888]: I1124 00:49:02.087372 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerStarted","Data":"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c"} Nov 24 00:49:03 crc kubenswrapper[4888]: I1124 00:49:03.103654 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerStarted","Data":"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314"} Nov 24 00:49:04 crc kubenswrapper[4888]: I1124 00:49:04.122665 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerStarted","Data":"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189"} Nov 24 00:49:05 crc kubenswrapper[4888]: I1124 00:49:05.147948 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerStarted","Data":"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba"} Nov 24 00:49:05 crc kubenswrapper[4888]: I1124 00:49:05.149010 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:49:05 crc kubenswrapper[4888]: I1124 00:49:05.184372 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.858858232 podStartE2EDuration="5.184348545s" podCreationTimestamp="2025-11-24 00:49:00 +0000 UTC" firstStartedPulling="2025-11-24 00:49:01.012550685 +0000 UTC m=+1443.595234729" lastFinishedPulling="2025-11-24 00:49:04.338040988 +0000 UTC m=+1446.920725042" observedRunningTime="2025-11-24 00:49:05.173971975 +0000 UTC m=+1447.756656019" watchObservedRunningTime="2025-11-24 00:49:05.184348545 +0000 UTC m=+1447.767032589" Nov 24 00:49:06 crc kubenswrapper[4888]: I1124 00:49:06.489570 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 00:49:23 crc kubenswrapper[4888]: I1124 00:49:23.592271 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:49:23 crc kubenswrapper[4888]: I1124 00:49:23.592885 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:49:30 crc kubenswrapper[4888]: I1124 00:49:30.491926 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.631529 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-2qlfb"] Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.647738 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-2qlfb"] Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.747804 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-cpbbm"] Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.755007 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.794273 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-cpbbm"] Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.880132 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.880623 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg2fh\" (UniqueName: \"kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.880781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.982905 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.983168 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg2fh\" (UniqueName: \"kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.983256 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.989518 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.996625 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:42 crc kubenswrapper[4888]: I1124 00:49:42.998798 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg2fh\" (UniqueName: \"kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh\") pod \"heat-db-sync-cpbbm\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:43 crc kubenswrapper[4888]: I1124 00:49:43.093646 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cpbbm" Nov 24 00:49:43 crc kubenswrapper[4888]: I1124 00:49:43.565958 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-cpbbm"] Nov 24 00:49:43 crc kubenswrapper[4888]: I1124 00:49:43.572267 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:49:43 crc kubenswrapper[4888]: I1124 00:49:43.670828 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cpbbm" event={"ID":"6877fe07-b619-4e4e-8b55-ed047924a852","Type":"ContainerStarted","Data":"d59c68caad525178956c8fae178e941d5cbafa732abae6b81072990d1e09f359"} Nov 24 00:49:44 crc kubenswrapper[4888]: I1124 00:49:44.261992 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5cc9373-09e2-4cae-82e0-badb0da1e0ad" path="/var/lib/kubelet/pods/f5cc9373-09e2-4cae-82e0-badb0da1e0ad/volumes" Nov 24 00:49:44 crc kubenswrapper[4888]: I1124 00:49:44.375167 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.000998 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.001327 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-central-agent" containerID="cri-o://4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c" gracePeriod=30 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.001432 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="sg-core" containerID="cri-o://1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189" gracePeriod=30 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.001409 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="proxy-httpd" containerID="cri-o://e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba" gracePeriod=30 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.001468 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-notification-agent" containerID="cri-o://64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314" gracePeriod=30 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.231841 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718087 4888 generic.go:334] "Generic (PLEG): container finished" podID="2a956701-1723-4960-8b2f-cd581eeb4855" containerID="e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba" exitCode=0 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718117 4888 generic.go:334] "Generic (PLEG): container finished" podID="2a956701-1723-4960-8b2f-cd581eeb4855" containerID="1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189" exitCode=2 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718126 4888 generic.go:334] "Generic (PLEG): container finished" podID="2a956701-1723-4960-8b2f-cd581eeb4855" containerID="4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c" exitCode=0 Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718154 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerDied","Data":"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba"} Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718179 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerDied","Data":"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189"} Nov 24 00:49:45 crc kubenswrapper[4888]: I1124 00:49:45.718190 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerDied","Data":"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c"} Nov 24 00:49:48 crc kubenswrapper[4888]: I1124 00:49:48.710222 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="rabbitmq" containerID="cri-o://f00812e0f8369a19c2f0bf3b384086717a2595fd39171ef4559011cc873d78e6" gracePeriod=604796 Nov 24 00:49:49 crc kubenswrapper[4888]: I1124 00:49:49.232668 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="rabbitmq" containerID="cri-o://ffa92dbfd86603ffbe9031bbd1601e583f01c66d3d29826cb69f3b20bbb67ec4" gracePeriod=604796 Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.305603 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.340604 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq827\" (UniqueName: \"kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.341133 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.348581 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827" (OuterVolumeSpecName: "kube-api-access-qq827") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "kube-api-access-qq827". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.419066 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.442917 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.443174 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.443326 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.443479 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.444151 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.444354 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd\") pod \"2a956701-1723-4960-8b2f-cd581eeb4855\" (UID: \"2a956701-1723-4960-8b2f-cd581eeb4855\") " Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.444231 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.444586 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.445067 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.445175 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq827\" (UniqueName: \"kubernetes.io/projected/2a956701-1723-4960-8b2f-cd581eeb4855-kube-api-access-qq827\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.445247 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.445331 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a956701-1723-4960-8b2f-cd581eeb4855-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.447958 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts" (OuterVolumeSpecName: "scripts") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.480846 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.540678 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.546748 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.546775 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.546786 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.561384 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data" (OuterVolumeSpecName: "config-data") pod "2a956701-1723-4960-8b2f-cd581eeb4855" (UID: "2a956701-1723-4960-8b2f-cd581eeb4855"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.647426 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a956701-1723-4960-8b2f-cd581eeb4855-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.769941 4888 generic.go:334] "Generic (PLEG): container finished" podID="2a956701-1723-4960-8b2f-cd581eeb4855" containerID="64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314" exitCode=0 Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.769989 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerDied","Data":"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314"} Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.770025 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a956701-1723-4960-8b2f-cd581eeb4855","Type":"ContainerDied","Data":"ccfd4e883d77908cdec2d69373f9d0b8c5f4a36e90b83266ffb88105e989131b"} Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.770041 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.770045 4888 scope.go:117] "RemoveContainer" containerID="e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.796121 4888 scope.go:117] "RemoveContainer" containerID="1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.805102 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.826670 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.834149 4888 scope.go:117] "RemoveContainer" containerID="64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843154 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.843670 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-central-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843688 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-central-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.843709 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="proxy-httpd" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843716 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="proxy-httpd" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.843744 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-notification-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843751 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-notification-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.843767 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="sg-core" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843773 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="sg-core" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.843996 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="sg-core" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.844014 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-central-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.844026 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="proxy-httpd" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.844037 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" containerName="ceilometer-notification-agent" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.846478 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849235 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849264 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849316 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849335 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849359 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849386 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849407 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.849454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h7j5\" (UniqueName: \"kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.850018 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.850834 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.853744 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.856067 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.877372 4888 scope.go:117] "RemoveContainer" containerID="4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.904131 4888 scope.go:117] "RemoveContainer" containerID="e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.904631 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba\": container with ID starting with e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba not found: ID does not exist" containerID="e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.904661 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba"} err="failed to get container status \"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba\": rpc error: code = NotFound desc = could not find container \"e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba\": container with ID starting with e299a31a4e2f22802029d6ff4ac4d9ba90e1565c465d36aeef27dc6aaed50cba not found: ID does not exist" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.904683 4888 scope.go:117] "RemoveContainer" containerID="1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.905292 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189\": container with ID starting with 1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189 not found: ID does not exist" containerID="1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.905318 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189"} err="failed to get container status \"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189\": rpc error: code = NotFound desc = could not find container \"1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189\": container with ID starting with 1f3b77355f3665c3924238f1d7ac44e32c70716ebe14c72ac84940c5ba36c189 not found: ID does not exist" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.905335 4888 scope.go:117] "RemoveContainer" containerID="64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.905744 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314\": container with ID starting with 64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314 not found: ID does not exist" containerID="64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.905770 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314"} err="failed to get container status \"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314\": rpc error: code = NotFound desc = could not find container \"64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314\": container with ID starting with 64b813abc84570c2855c3d2bba05903e7f9c89521cf57d05de2440e31b921314 not found: ID does not exist" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.905783 4888 scope.go:117] "RemoveContainer" containerID="4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c" Nov 24 00:49:50 crc kubenswrapper[4888]: E1124 00:49:50.906252 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c\": container with ID starting with 4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c not found: ID does not exist" containerID="4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.906292 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c"} err="failed to get container status \"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c\": rpc error: code = NotFound desc = could not find container \"4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c\": container with ID starting with 4e9f86e458b456fd552fc55632d00d73c36b76f6bdc6d68a441cf10b8debbe5c not found: ID does not exist" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950367 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h7j5\" (UniqueName: \"kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950472 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950490 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950543 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950562 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950585 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950614 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.950634 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.951393 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.953385 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.960920 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.961135 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.962946 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.964750 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.974592 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:50 crc kubenswrapper[4888]: I1124 00:49:50.976585 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h7j5\" (UniqueName: \"kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5\") pod \"ceilometer-0\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " pod="openstack/ceilometer-0" Nov 24 00:49:51 crc kubenswrapper[4888]: I1124 00:49:51.173642 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 00:49:51 crc kubenswrapper[4888]: I1124 00:49:51.677574 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 00:49:51 crc kubenswrapper[4888]: I1124 00:49:51.790291 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerStarted","Data":"c02f9cfdeb7bbf6a4b4ffb6eb9662e2922098ccedb060ecff4541896b33d7a17"} Nov 24 00:49:52 crc kubenswrapper[4888]: I1124 00:49:52.264241 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a956701-1723-4960-8b2f-cd581eeb4855" path="/var/lib/kubelet/pods/2a956701-1723-4960-8b2f-cd581eeb4855/volumes" Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.591480 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.591762 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.591802 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.592564 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.592619 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" gracePeriod=600 Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.816313 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" exitCode=0 Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.816386 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153"} Nov 24 00:49:53 crc kubenswrapper[4888]: I1124 00:49:53.816454 4888 scope.go:117] "RemoveContainer" containerID="574856c5292a1cdd825286b173bd752aa267e9d62d42f7843f7e2afdf744889c" Nov 24 00:49:55 crc kubenswrapper[4888]: I1124 00:49:55.848913 4888 generic.go:334] "Generic (PLEG): container finished" podID="19077405-4b37-4747-a4aa-37013176475e" containerID="ffa92dbfd86603ffbe9031bbd1601e583f01c66d3d29826cb69f3b20bbb67ec4" exitCode=0 Nov 24 00:49:55 crc kubenswrapper[4888]: I1124 00:49:55.849114 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerDied","Data":"ffa92dbfd86603ffbe9031bbd1601e583f01c66d3d29826cb69f3b20bbb67ec4"} Nov 24 00:49:55 crc kubenswrapper[4888]: I1124 00:49:55.852728 4888 generic.go:334] "Generic (PLEG): container finished" podID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerID="f00812e0f8369a19c2f0bf3b384086717a2595fd39171ef4559011cc873d78e6" exitCode=0 Nov 24 00:49:55 crc kubenswrapper[4888]: I1124 00:49:55.852774 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerDied","Data":"f00812e0f8369a19c2f0bf3b384086717a2595fd39171ef4559011cc873d78e6"} Nov 24 00:49:57 crc kubenswrapper[4888]: E1124 00:49:57.084561 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:49:57 crc kubenswrapper[4888]: I1124 00:49:57.877002 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:49:57 crc kubenswrapper[4888]: E1124 00:49:57.877998 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.758849 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.760738 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.764012 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779698 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779741 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779770 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779797 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbw4d\" (UniqueName: \"kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779872 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779911 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.779931 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.793406 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881521 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881599 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881624 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881707 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881735 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881763 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.881790 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbw4d\" (UniqueName: \"kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.882448 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.882603 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.888147 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.890218 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.890791 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.890895 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:49:59 crc kubenswrapper[4888]: I1124 00:49:59.912081 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbw4d\" (UniqueName: \"kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d\") pod \"dnsmasq-dns-7d84b4d45c-pjxgn\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:00 crc kubenswrapper[4888]: I1124 00:50:00.083100 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:02 crc kubenswrapper[4888]: I1124 00:50:02.624853 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.121:5671: i/o timeout" Nov 24 00:50:02 crc kubenswrapper[4888]: I1124 00:50:02.923164 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.122:5671: i/o timeout" Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.962900 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c51b1de1-34e6-48a3-8950-cdfaf80d28a0","Type":"ContainerDied","Data":"8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42"} Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.963900 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c50ca3995e28592a0d5189f02254001308b7cfdfeff878c664e6a06a06eea42" Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.965775 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"19077405-4b37-4747-a4aa-37013176475e","Type":"ContainerDied","Data":"60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf"} Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.965799 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60e5a1bc73c44fbfa0b62ba3366678eaeb302b18d6457e57d2fa5d370862e8cf" Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.985732 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:50:04 crc kubenswrapper[4888]: I1124 00:50:04.994146 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.094787 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.094923 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.094978 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095082 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095133 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095162 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9chg\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095214 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095289 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095335 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095396 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095460 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095520 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095555 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc5dl\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095581 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095620 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095643 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095671 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095725 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095750 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095788 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.098842 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.095859 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data\") pod \"19077405-4b37-4747-a4aa-37013176475e\" (UID: \"19077405-4b37-4747-a4aa-37013176475e\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.099164 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info\") pod \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\" (UID: \"c51b1de1-34e6-48a3-8950-cdfaf80d28a0\") " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.100133 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.100324 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.106666 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.107087 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.107714 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.109247 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.109565 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.115572 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg" (OuterVolumeSpecName: "kube-api-access-p9chg") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "kube-api-access-p9chg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.116544 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info" (OuterVolumeSpecName: "pod-info") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.120149 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.124990 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info" (OuterVolumeSpecName: "pod-info") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.130399 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.132523 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.137258 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.138004 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.149457 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl" (OuterVolumeSpecName: "kube-api-access-hc5dl") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "kube-api-access-hc5dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.180279 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data" (OuterVolumeSpecName: "config-data") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204225 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc5dl\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-kube-api-access-hc5dl\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204263 4888 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204277 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204289 4888 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19077405-4b37-4747-a4aa-37013176475e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204320 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204332 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204343 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204353 4888 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204364 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204373 4888 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19077405-4b37-4747-a4aa-37013176475e-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204385 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204398 4888 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204410 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9chg\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-kube-api-access-p9chg\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204422 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204438 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.204448 4888 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.263186 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf" (OuterVolumeSpecName: "server-conf") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.278240 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.311250 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data" (OuterVolumeSpecName: "config-data") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.311750 4888 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.311780 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.311788 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19077405-4b37-4747-a4aa-37013176475e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.321890 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.362866 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf" (OuterVolumeSpecName: "server-conf") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.377036 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c51b1de1-34e6-48a3-8950-cdfaf80d28a0" (UID: "c51b1de1-34e6-48a3-8950-cdfaf80d28a0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.377400 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "19077405-4b37-4747-a4aa-37013176475e" (UID: "19077405-4b37-4747-a4aa-37013176475e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.413497 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19077405-4b37-4747-a4aa-37013176475e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.413535 4888 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.413546 4888 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c51b1de1-34e6-48a3-8950-cdfaf80d28a0-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: I1124 00:50:05.413559 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:05 crc kubenswrapper[4888]: E1124 00:50:05.650836 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 24 00:50:05 crc kubenswrapper[4888]: E1124 00:50:05.651087 4888 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 24 00:50:05 crc kubenswrapper[4888]: E1124 00:50:05.651211 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tg2fh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-cpbbm_openstack(6877fe07-b619-4e4e-8b55-ed047924a852): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 00:50:05 crc kubenswrapper[4888]: E1124 00:50:05.652361 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-cpbbm" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.000585 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.000585 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: E1124 00:50:06.004849 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-cpbbm" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.102882 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.131082 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.148644 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.170900 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214172 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: E1124 00:50:06.214667 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214682 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: E1124 00:50:06.214697 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="setup-container" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214703 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="setup-container" Nov 24 00:50:06 crc kubenswrapper[4888]: E1124 00:50:06.214730 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214735 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: E1124 00:50:06.214749 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="setup-container" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214756 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="setup-container" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.214975 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.215000 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="19077405-4b37-4747-a4aa-37013176475e" containerName="rabbitmq" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.217070 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.224181 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.229348 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.229633 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.229746 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.229934 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.230035 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-gqh27" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.230641 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.238874 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.240763 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.246888 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.247068 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.247182 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.247314 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.247822 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.248094 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.248206 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lvpwf" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.284528 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19077405-4b37-4747-a4aa-37013176475e" path="/var/lib/kubelet/pods/19077405-4b37-4747-a4aa-37013176475e/volumes" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.287079 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51b1de1-34e6-48a3-8950-cdfaf80d28a0" path="/var/lib/kubelet/pods/c51b1de1-34e6-48a3-8950-cdfaf80d28a0/volumes" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.289673 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.292755 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334732 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334783 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-config-data\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334827 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334847 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334895 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334921 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.334948 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335112 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335135 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335163 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/021b7361-39f2-47fa-b028-b1656b5fe94c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335200 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f987cf1a-247e-411e-9d03-4a5f44a98d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335274 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/021b7361-39f2-47fa-b028-b1656b5fe94c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335303 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q72k\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-kube-api-access-5q72k\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335327 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335484 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335530 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335615 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mwk9\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-kube-api-access-4mwk9\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335712 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335742 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f987cf1a-247e-411e-9d03-4a5f44a98d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335772 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.335795 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438270 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438352 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438371 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438399 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mwk9\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-kube-api-access-4mwk9\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438429 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f987cf1a-247e-411e-9d03-4a5f44a98d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438463 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438479 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438504 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438524 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-config-data\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438542 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438560 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438589 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438608 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438635 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438669 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438685 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438711 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/021b7361-39f2-47fa-b028-b1656b5fe94c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438729 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438748 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f987cf1a-247e-411e-9d03-4a5f44a98d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438796 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/021b7361-39f2-47fa-b028-b1656b5fe94c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.438828 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q72k\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-kube-api-access-5q72k\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.439339 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.440013 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.440623 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.440739 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.442536 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.441045 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-config-data\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.441976 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.442083 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.442230 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/021b7361-39f2-47fa-b028-b1656b5fe94c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.442425 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f987cf1a-247e-411e-9d03-4a5f44a98d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.441762 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.447672 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.447797 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.452886 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/021b7361-39f2-47fa-b028-b1656b5fe94c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.453166 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.454364 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f987cf1a-247e-411e-9d03-4a5f44a98d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.455202 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f987cf1a-247e-411e-9d03-4a5f44a98d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.455248 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f987cf1a-247e-411e-9d03-4a5f44a98d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.455480 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/021b7361-39f2-47fa-b028-b1656b5fe94c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.457087 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.464143 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mwk9\" (UniqueName: \"kubernetes.io/projected/021b7361-39f2-47fa-b028-b1656b5fe94c-kube-api-access-4mwk9\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.467784 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q72k\" (UniqueName: \"kubernetes.io/projected/f987cf1a-247e-411e-9d03-4a5f44a98d14-kube-api-access-5q72k\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.484879 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f987cf1a-247e-411e-9d03-4a5f44a98d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.504391 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"021b7361-39f2-47fa-b028-b1656b5fe94c\") " pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.557983 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 00:50:06 crc kubenswrapper[4888]: I1124 00:50:06.577333 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:07 crc kubenswrapper[4888]: I1124 00:50:07.452066 4888 scope.go:117] "RemoveContainer" containerID="77f92ef1d28bc7ff0790ff6bf0779443bddf3af848718975cf6a1958e090c6bb" Nov 24 00:50:08 crc kubenswrapper[4888]: I1124 00:50:08.006536 4888 scope.go:117] "RemoveContainer" containerID="d49093346fd5a6787c95e5e78049f40011ec54168f4dcdf7c2296c2c4d02c337" Nov 24 00:50:08 crc kubenswrapper[4888]: I1124 00:50:08.432112 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:50:08 crc kubenswrapper[4888]: I1124 00:50:08.668923 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 00:50:08 crc kubenswrapper[4888]: I1124 00:50:08.680113 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.050518 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerStarted","Data":"587ed132ea34f3ebadc888b2b3c0e909483c9272d0497faeded2b6ebb3fc128b"} Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.052333 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f987cf1a-247e-411e-9d03-4a5f44a98d14","Type":"ContainerStarted","Data":"e8e7760afbd493510eb74155e0fd2b477ed0ff72e28df7873269c7ee90e47077"} Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.054129 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"021b7361-39f2-47fa-b028-b1656b5fe94c","Type":"ContainerStarted","Data":"5b9a9676f129e74576cc5c59810e7a4da6355574cf1f12815f979c7f22180991"} Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.056276 4888 generic.go:334] "Generic (PLEG): container finished" podID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerID="5c0e564b20831b48d8d11040a6153ab064106af2e2aa09fb44949bdc00a20cf5" exitCode=0 Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.056315 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" event={"ID":"da207a65-f598-465d-bb3b-1cf89c4ee003","Type":"ContainerDied","Data":"5c0e564b20831b48d8d11040a6153ab064106af2e2aa09fb44949bdc00a20cf5"} Nov 24 00:50:09 crc kubenswrapper[4888]: I1124 00:50:09.056341 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" event={"ID":"da207a65-f598-465d-bb3b-1cf89c4ee003","Type":"ContainerStarted","Data":"3784760548fd5ceaf98482c5462e3b0052705920ba0945a4709ef64282f0b6b9"} Nov 24 00:50:10 crc kubenswrapper[4888]: I1124 00:50:10.077879 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerStarted","Data":"6aba00bfee962dd890d93a5867f963b103ae0c8dbc9c95abcc626cdcd747548e"} Nov 24 00:50:10 crc kubenswrapper[4888]: I1124 00:50:10.081669 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" event={"ID":"da207a65-f598-465d-bb3b-1cf89c4ee003","Type":"ContainerStarted","Data":"c7d988d707d7809feb4c2c9b1c87db8a0bdbd76a81c53f34782f3cb32996b564"} Nov 24 00:50:10 crc kubenswrapper[4888]: I1124 00:50:10.081847 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:10 crc kubenswrapper[4888]: I1124 00:50:10.136306 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" podStartSLOduration=11.136288846 podStartE2EDuration="11.136288846s" podCreationTimestamp="2025-11-24 00:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:50:10.123742063 +0000 UTC m=+1512.706426107" watchObservedRunningTime="2025-11-24 00:50:10.136288846 +0000 UTC m=+1512.718972890" Nov 24 00:50:11 crc kubenswrapper[4888]: I1124 00:50:11.096345 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f987cf1a-247e-411e-9d03-4a5f44a98d14","Type":"ContainerStarted","Data":"a8207573f74db48f974439d2e7b567d467fff83424b51d42ad0cdfc3ec5618ce"} Nov 24 00:50:11 crc kubenswrapper[4888]: I1124 00:50:11.098923 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"021b7361-39f2-47fa-b028-b1656b5fe94c","Type":"ContainerStarted","Data":"eea65ea1d457f5113f5c553d1b73980bfc751f5f5d703990dbb2ec5aeef9c994"} Nov 24 00:50:11 crc kubenswrapper[4888]: I1124 00:50:11.105600 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerStarted","Data":"02d1227a7ec42fa3f8790c2eb85236a26e280b2afc33ed698fcc045e07cb6fe7"} Nov 24 00:50:13 crc kubenswrapper[4888]: I1124 00:50:13.124524 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerStarted","Data":"d7c28ede11c7103214633efc6e8d69d4b1af684a72e184c5f75ccbcb8349256d"} Nov 24 00:50:13 crc kubenswrapper[4888]: I1124 00:50:13.125169 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 00:50:13 crc kubenswrapper[4888]: I1124 00:50:13.144534 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.95086855 podStartE2EDuration="23.144513799s" podCreationTimestamp="2025-11-24 00:49:50 +0000 UTC" firstStartedPulling="2025-11-24 00:49:51.690203593 +0000 UTC m=+1494.272887637" lastFinishedPulling="2025-11-24 00:50:11.883848832 +0000 UTC m=+1514.466532886" observedRunningTime="2025-11-24 00:50:13.143165321 +0000 UTC m=+1515.725849365" watchObservedRunningTime="2025-11-24 00:50:13.144513799 +0000 UTC m=+1515.727197853" Nov 24 00:50:13 crc kubenswrapper[4888]: I1124 00:50:13.246017 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:50:13 crc kubenswrapper[4888]: E1124 00:50:13.246388 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.085080 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.188272 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.188491 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="dnsmasq-dns" containerID="cri-o://99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4" gracePeriod=10 Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.350331 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.352366 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.368505 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.433579 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.433663 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.433700 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.434016 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg6rm\" (UniqueName: \"kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.434286 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.434363 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.434412 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.536585 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.536663 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.536689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.536720 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg6rm\" (UniqueName: \"kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.537088 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.537130 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.537161 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.537803 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.537856 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.538260 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.538284 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.538570 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.587691 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.589972 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg6rm\" (UniqueName: \"kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm\") pod \"dnsmasq-dns-6559847fc9-m5kg8\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.679544 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.881479 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953314 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953384 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953456 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953478 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953609 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.953643 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kdr6\" (UniqueName: \"kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6\") pod \"d115b693-6451-46a1-a81e-327eea577d2b\" (UID: \"d115b693-6451-46a1-a81e-327eea577d2b\") " Nov 24 00:50:15 crc kubenswrapper[4888]: I1124 00:50:15.959003 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6" (OuterVolumeSpecName: "kube-api-access-6kdr6") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "kube-api-access-6kdr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.038568 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config" (OuterVolumeSpecName: "config") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.038621 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.046201 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.056170 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.056198 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.056208 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.056219 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kdr6\" (UniqueName: \"kubernetes.io/projected/d115b693-6451-46a1-a81e-327eea577d2b-kube-api-access-6kdr6\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.060949 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.061662 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d115b693-6451-46a1-a81e-327eea577d2b" (UID: "d115b693-6451-46a1-a81e-327eea577d2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.157754 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.157779 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d115b693-6451-46a1-a81e-327eea577d2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.160632 4888 generic.go:334] "Generic (PLEG): container finished" podID="d115b693-6451-46a1-a81e-327eea577d2b" containerID="99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4" exitCode=0 Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.160664 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" event={"ID":"d115b693-6451-46a1-a81e-327eea577d2b","Type":"ContainerDied","Data":"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4"} Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.160689 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" event={"ID":"d115b693-6451-46a1-a81e-327eea577d2b","Type":"ContainerDied","Data":"c21a92cd7079c47491d8e1c60edec8ac73f74992f8722ccd43ada454a84642c4"} Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.160706 4888 scope.go:117] "RemoveContainer" containerID="99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.160861 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.194369 4888 scope.go:117] "RemoveContainer" containerID="ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.201866 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.210403 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kz8ll"] Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.225634 4888 scope.go:117] "RemoveContainer" containerID="99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4" Nov 24 00:50:16 crc kubenswrapper[4888]: E1124 00:50:16.226932 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4\": container with ID starting with 99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4 not found: ID does not exist" containerID="99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.226968 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4"} err="failed to get container status \"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4\": rpc error: code = NotFound desc = could not find container \"99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4\": container with ID starting with 99e22e99046cc39eeb7a665467a096262c01971376df74f7d0f4ffc8439d7cf4 not found: ID does not exist" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.226999 4888 scope.go:117] "RemoveContainer" containerID="ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca" Nov 24 00:50:16 crc kubenswrapper[4888]: E1124 00:50:16.227296 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca\": container with ID starting with ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca not found: ID does not exist" containerID="ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.227365 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca"} err="failed to get container status \"ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca\": rpc error: code = NotFound desc = could not find container \"ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca\": container with ID starting with ef603f200cb49368254572a91565b7aecf64a215fa3607537f9d075c336027ca not found: ID does not exist" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.262298 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d115b693-6451-46a1-a81e-327eea577d2b" path="/var/lib/kubelet/pods/d115b693-6451-46a1-a81e-327eea577d2b/volumes" Nov 24 00:50:16 crc kubenswrapper[4888]: I1124 00:50:16.292582 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 00:50:17 crc kubenswrapper[4888]: I1124 00:50:17.185172 4888 generic.go:334] "Generic (PLEG): container finished" podID="377504e9-345b-4365-a0b4-3922c341e4d9" containerID="ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a" exitCode=0 Nov 24 00:50:17 crc kubenswrapper[4888]: I1124 00:50:17.185238 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" event={"ID":"377504e9-345b-4365-a0b4-3922c341e4d9","Type":"ContainerDied","Data":"ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a"} Nov 24 00:50:17 crc kubenswrapper[4888]: I1124 00:50:17.185505 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" event={"ID":"377504e9-345b-4365-a0b4-3922c341e4d9","Type":"ContainerStarted","Data":"76e5183791a01f810c373f3ce6f270b772fece3df443563cb92f7d33c51aca19"} Nov 24 00:50:18 crc kubenswrapper[4888]: I1124 00:50:18.208655 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" event={"ID":"377504e9-345b-4365-a0b4-3922c341e4d9","Type":"ContainerStarted","Data":"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720"} Nov 24 00:50:18 crc kubenswrapper[4888]: I1124 00:50:18.209043 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:18 crc kubenswrapper[4888]: I1124 00:50:18.247784 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" podStartSLOduration=3.247749964 podStartE2EDuration="3.247749964s" podCreationTimestamp="2025-11-24 00:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:50:18.238173315 +0000 UTC m=+1520.820857389" watchObservedRunningTime="2025-11-24 00:50:18.247749964 +0000 UTC m=+1520.830434048" Nov 24 00:50:20 crc kubenswrapper[4888]: I1124 00:50:20.235629 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cpbbm" event={"ID":"6877fe07-b619-4e4e-8b55-ed047924a852","Type":"ContainerStarted","Data":"6cadd228fdd7a0a141c0eb92bd0de6b2524d3a9f1b97fad58422fc078f1eb641"} Nov 24 00:50:20 crc kubenswrapper[4888]: I1124 00:50:20.267107 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-cpbbm" podStartSLOduration=2.382341845 podStartE2EDuration="38.267091876s" podCreationTimestamp="2025-11-24 00:49:42 +0000 UTC" firstStartedPulling="2025-11-24 00:49:43.571572065 +0000 UTC m=+1486.154256119" lastFinishedPulling="2025-11-24 00:50:19.456322096 +0000 UTC m=+1522.039006150" observedRunningTime="2025-11-24 00:50:20.25978512 +0000 UTC m=+1522.842469164" watchObservedRunningTime="2025-11-24 00:50:20.267091876 +0000 UTC m=+1522.849775920" Nov 24 00:50:23 crc kubenswrapper[4888]: I1124 00:50:23.275800 4888 generic.go:334] "Generic (PLEG): container finished" podID="6877fe07-b619-4e4e-8b55-ed047924a852" containerID="6cadd228fdd7a0a141c0eb92bd0de6b2524d3a9f1b97fad58422fc078f1eb641" exitCode=0 Nov 24 00:50:23 crc kubenswrapper[4888]: I1124 00:50:23.275908 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cpbbm" event={"ID":"6877fe07-b619-4e4e-8b55-ed047924a852","Type":"ContainerDied","Data":"6cadd228fdd7a0a141c0eb92bd0de6b2524d3a9f1b97fad58422fc078f1eb641"} Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.244921 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:50:24 crc kubenswrapper[4888]: E1124 00:50:24.245610 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.732009 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cpbbm" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.804183 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle\") pod \"6877fe07-b619-4e4e-8b55-ed047924a852\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.804446 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg2fh\" (UniqueName: \"kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh\") pod \"6877fe07-b619-4e4e-8b55-ed047924a852\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.804556 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data\") pod \"6877fe07-b619-4e4e-8b55-ed047924a852\" (UID: \"6877fe07-b619-4e4e-8b55-ed047924a852\") " Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.817077 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh" (OuterVolumeSpecName: "kube-api-access-tg2fh") pod "6877fe07-b619-4e4e-8b55-ed047924a852" (UID: "6877fe07-b619-4e4e-8b55-ed047924a852"). InnerVolumeSpecName "kube-api-access-tg2fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.843015 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6877fe07-b619-4e4e-8b55-ed047924a852" (UID: "6877fe07-b619-4e4e-8b55-ed047924a852"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.893699 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data" (OuterVolumeSpecName: "config-data") pod "6877fe07-b619-4e4e-8b55-ed047924a852" (UID: "6877fe07-b619-4e4e-8b55-ed047924a852"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.906840 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.907019 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877fe07-b619-4e4e-8b55-ed047924a852-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:24 crc kubenswrapper[4888]: I1124 00:50:24.907080 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg2fh\" (UniqueName: \"kubernetes.io/projected/6877fe07-b619-4e4e-8b55-ed047924a852-kube-api-access-tg2fh\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.307801 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cpbbm" event={"ID":"6877fe07-b619-4e4e-8b55-ed047924a852","Type":"ContainerDied","Data":"d59c68caad525178956c8fae178e941d5cbafa732abae6b81072990d1e09f359"} Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.308188 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d59c68caad525178956c8fae178e941d5cbafa732abae6b81072990d1e09f359" Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.307914 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cpbbm" Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.681913 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.749501 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:50:25 crc kubenswrapper[4888]: I1124 00:50:25.749770 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="dnsmasq-dns" containerID="cri-o://c7d988d707d7809feb4c2c9b1c87db8a0bdbd76a81c53f34782f3cb32996b564" gracePeriod=10 Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.318003 4888 generic.go:334] "Generic (PLEG): container finished" podID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerID="c7d988d707d7809feb4c2c9b1c87db8a0bdbd76a81c53f34782f3cb32996b564" exitCode=0 Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.318017 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" event={"ID":"da207a65-f598-465d-bb3b-1cf89c4ee003","Type":"ContainerDied","Data":"c7d988d707d7809feb4c2c9b1c87db8a0bdbd76a81c53f34782f3cb32996b564"} Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.790461 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839078 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-799c49dfcf-w6lnt"] Nov 24 00:50:26 crc kubenswrapper[4888]: E1124 00:50:26.839546 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="init" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839561 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="init" Nov 24 00:50:26 crc kubenswrapper[4888]: E1124 00:50:26.839571 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" containerName="heat-db-sync" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839578 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" containerName="heat-db-sync" Nov 24 00:50:26 crc kubenswrapper[4888]: E1124 00:50:26.839626 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839636 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: E1124 00:50:26.839661 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839668 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: E1124 00:50:26.839677 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="init" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839683 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="init" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839942 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="d115b693-6451-46a1-a81e-327eea577d2b" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839971 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" containerName="dnsmasq-dns" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.839992 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" containerName="heat-db-sync" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.840937 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855055 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855125 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855152 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855275 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbw4d\" (UniqueName: \"kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855405 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855445 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.855475 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config\") pod \"da207a65-f598-465d-bb3b-1cf89c4ee003\" (UID: \"da207a65-f598-465d-bb3b-1cf89c4ee003\") " Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.890154 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-799c49dfcf-w6lnt"] Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.921066 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d" (OuterVolumeSpecName: "kube-api-access-qbw4d") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "kube-api-access-qbw4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.971254 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2t2\" (UniqueName: \"kubernetes.io/projected/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-kube-api-access-tx2t2\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.973096 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data-custom\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.973526 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.973837 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-combined-ca-bundle\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:26 crc kubenswrapper[4888]: I1124 00:50:26.974234 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbw4d\" (UniqueName: \"kubernetes.io/projected/da207a65-f598-465d-bb3b-1cf89c4ee003-kube-api-access-qbw4d\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.047093 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7f5f5cb65-zfpq5"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.048537 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.062356 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f5f5cb65-zfpq5"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.079200 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082145 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2t2\" (UniqueName: \"kubernetes.io/projected/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-kube-api-access-tx2t2\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082334 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data-custom\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082467 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082550 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-internal-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082648 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.082733 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxcvj\" (UniqueName: \"kubernetes.io/projected/6baa14da-0b8b-43da-81b9-605e596cbb11-kube-api-access-zxcvj\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.083783 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-combined-ca-bundle\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.084034 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-combined-ca-bundle\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.084173 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-public-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.084343 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data-custom\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.084490 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.090240 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data-custom\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.101262 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-config-data\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.103781 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-775b95987-88pxg"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.104365 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.105397 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.110067 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.119325 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-775b95987-88pxg"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.119387 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2t2\" (UniqueName: \"kubernetes.io/projected/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-kube-api-access-tx2t2\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.127542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef846e73-7a80-4d25-ac41-cfa95d68cc1e-combined-ca-bundle\") pod \"heat-engine-799c49dfcf-w6lnt\" (UID: \"ef846e73-7a80-4d25-ac41-cfa95d68cc1e\") " pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.138264 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.142156 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config" (OuterVolumeSpecName: "config") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.149167 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "da207a65-f598-465d-bb3b-1cf89c4ee003" (UID: "da207a65-f598-465d-bb3b-1cf89c4ee003"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.179077 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185440 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-public-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185474 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-internal-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185493 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185521 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data-custom\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185541 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxcvj\" (UniqueName: \"kubernetes.io/projected/6baa14da-0b8b-43da-81b9-605e596cbb11-kube-api-access-zxcvj\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185691 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-combined-ca-bundle\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185728 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-combined-ca-bundle\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185752 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-internal-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185774 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-public-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185819 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185841 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data-custom\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185872 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746fq\" (UniqueName: \"kubernetes.io/projected/eb06d977-15f5-4263-aa7e-4fa2b2964847-kube-api-access-746fq\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185921 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185932 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185941 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185951 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.185960 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da207a65-f598-465d-bb3b-1cf89c4ee003-config\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.188994 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-internal-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.189768 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-combined-ca-bundle\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.190603 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-public-tls-certs\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.193572 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.193852 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6baa14da-0b8b-43da-81b9-605e596cbb11-config-data-custom\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.205248 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxcvj\" (UniqueName: \"kubernetes.io/projected/6baa14da-0b8b-43da-81b9-605e596cbb11-kube-api-access-zxcvj\") pod \"heat-api-7f5f5cb65-zfpq5\" (UID: \"6baa14da-0b8b-43da-81b9-605e596cbb11\") " pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.289561 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-combined-ca-bundle\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.289880 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-internal-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.289944 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.290028 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-746fq\" (UniqueName: \"kubernetes.io/projected/eb06d977-15f5-4263-aa7e-4fa2b2964847-kube-api-access-746fq\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.290105 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-public-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.290178 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data-custom\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.293904 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-internal-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.294952 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data-custom\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.296401 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-public-tls-certs\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.296403 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-config-data\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.299114 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb06d977-15f5-4263-aa7e-4fa2b2964847-combined-ca-bundle\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.307645 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-746fq\" (UniqueName: \"kubernetes.io/projected/eb06d977-15f5-4263-aa7e-4fa2b2964847-kube-api-access-746fq\") pod \"heat-cfnapi-775b95987-88pxg\" (UID: \"eb06d977-15f5-4263-aa7e-4fa2b2964847\") " pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.353170 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" event={"ID":"da207a65-f598-465d-bb3b-1cf89c4ee003","Type":"ContainerDied","Data":"3784760548fd5ceaf98482c5462e3b0052705920ba0945a4709ef64282f0b6b9"} Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.353262 4888 scope.go:117] "RemoveContainer" containerID="c7d988d707d7809feb4c2c9b1c87db8a0bdbd76a81c53f34782f3cb32996b564" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.353502 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-pjxgn" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.402306 4888 scope.go:117] "RemoveContainer" containerID="5c0e564b20831b48d8d11040a6153ab064106af2e2aa09fb44949bdc00a20cf5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.407414 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.416100 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-pjxgn"] Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.498350 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.504558 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:27 crc kubenswrapper[4888]: I1124 00:50:27.682284 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-799c49dfcf-w6lnt"] Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.033049 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-775b95987-88pxg"] Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.051776 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f5f5cb65-zfpq5"] Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.268226 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da207a65-f598-465d-bb3b-1cf89c4ee003" path="/var/lib/kubelet/pods/da207a65-f598-465d-bb3b-1cf89c4ee003/volumes" Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.382999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f5f5cb65-zfpq5" event={"ID":"6baa14da-0b8b-43da-81b9-605e596cbb11","Type":"ContainerStarted","Data":"b283d59ed0cd4b21be578fda2f4ceef4f5f237b9de709e25cc18aef48052da5b"} Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.389139 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-775b95987-88pxg" event={"ID":"eb06d977-15f5-4263-aa7e-4fa2b2964847","Type":"ContainerStarted","Data":"8bdd9f56fe2a15cd91ef832313c3c36e2aa71cda10ec043eaa19d7af9040fd09"} Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.395053 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-799c49dfcf-w6lnt" event={"ID":"ef846e73-7a80-4d25-ac41-cfa95d68cc1e","Type":"ContainerStarted","Data":"c9b44406085c6841809208a591403fa596ebd799c7a6089e365df88ae6ed4f4e"} Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.395101 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-799c49dfcf-w6lnt" event={"ID":"ef846e73-7a80-4d25-ac41-cfa95d68cc1e","Type":"ContainerStarted","Data":"8e137cea5a5a8aa27f107467b8782c86dc29d53ef3e48be7b7c0e1e6a24aa232"} Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.395148 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:28 crc kubenswrapper[4888]: I1124 00:50:28.423667 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-799c49dfcf-w6lnt" podStartSLOduration=2.423649793 podStartE2EDuration="2.423649793s" podCreationTimestamp="2025-11-24 00:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:50:28.417598493 +0000 UTC m=+1531.000282537" watchObservedRunningTime="2025-11-24 00:50:28.423649793 +0000 UTC m=+1531.006333837" Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.425874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-775b95987-88pxg" event={"ID":"eb06d977-15f5-4263-aa7e-4fa2b2964847","Type":"ContainerStarted","Data":"19fb6c245c2bc2393515e38f9a7473262d3f44cfa9d00efa22ef51f485c7b697"} Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.426783 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.428249 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f5f5cb65-zfpq5" event={"ID":"6baa14da-0b8b-43da-81b9-605e596cbb11","Type":"ContainerStarted","Data":"3f9fdfdd1bbfcca72a811bd57d6d23f3975ba970c5d183f723a5c4265233e169"} Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.431110 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.496336 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7f5f5cb65-zfpq5" podStartSLOduration=2.847466566 podStartE2EDuration="4.496320465s" podCreationTimestamp="2025-11-24 00:50:26 +0000 UTC" firstStartedPulling="2025-11-24 00:50:28.049139607 +0000 UTC m=+1530.631823651" lastFinishedPulling="2025-11-24 00:50:29.697993506 +0000 UTC m=+1532.280677550" observedRunningTime="2025-11-24 00:50:30.493257299 +0000 UTC m=+1533.075941343" watchObservedRunningTime="2025-11-24 00:50:30.496320465 +0000 UTC m=+1533.079004509" Nov 24 00:50:30 crc kubenswrapper[4888]: I1124 00:50:30.501453 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-775b95987-88pxg" podStartSLOduration=2.850290476 podStartE2EDuration="4.501432249s" podCreationTimestamp="2025-11-24 00:50:26 +0000 UTC" firstStartedPulling="2025-11-24 00:50:28.041004638 +0000 UTC m=+1530.623688682" lastFinishedPulling="2025-11-24 00:50:29.692146411 +0000 UTC m=+1532.274830455" observedRunningTime="2025-11-24 00:50:30.476012894 +0000 UTC m=+1533.058696938" watchObservedRunningTime="2025-11-24 00:50:30.501432249 +0000 UTC m=+1533.084116293" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.871478 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs"] Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.875321 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.878685 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.879415 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.881029 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.881312 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.893524 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs"] Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.981287 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.981385 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd42s\" (UniqueName: \"kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.981527 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:33 crc kubenswrapper[4888]: I1124 00:50:33.981609 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.083544 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.083613 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd42s\" (UniqueName: \"kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.083660 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.083700 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.090053 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.090178 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.093112 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.109946 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd42s\" (UniqueName: \"kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.207657 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:34 crc kubenswrapper[4888]: I1124 00:50:34.852701 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs"] Nov 24 00:50:34 crc kubenswrapper[4888]: W1124 00:50:34.857665 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64895ae8_ff60_4598_b404_cdd6c14c0c02.slice/crio-5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad WatchSource:0}: Error finding container 5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad: Status 404 returned error can't find the container with id 5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad Nov 24 00:50:35 crc kubenswrapper[4888]: I1124 00:50:35.245911 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:50:35 crc kubenswrapper[4888]: E1124 00:50:35.246166 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:50:35 crc kubenswrapper[4888]: I1124 00:50:35.478425 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" event={"ID":"64895ae8-ff60-4598-b404-cdd6c14c0c02","Type":"ContainerStarted","Data":"5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad"} Nov 24 00:50:37 crc kubenswrapper[4888]: I1124 00:50:37.221345 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-799c49dfcf-w6lnt" Nov 24 00:50:37 crc kubenswrapper[4888]: I1124 00:50:37.283936 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:50:37 crc kubenswrapper[4888]: I1124 00:50:37.284167 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-595c95d7db-65z8t" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" containerID="cri-o://cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" gracePeriod=60 Nov 24 00:50:38 crc kubenswrapper[4888]: I1124 00:50:38.923189 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7f5f5cb65-zfpq5" Nov 24 00:50:38 crc kubenswrapper[4888]: I1124 00:50:38.983606 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:50:38 crc kubenswrapper[4888]: I1124 00:50:38.983801 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-69d5b6b95b-jhbt6" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerName="heat-api" containerID="cri-o://e3b1c60baf33e1c69193901f5b70d3bd167c112a9db60773ac0f39f047c8f838" gracePeriod=60 Nov 24 00:50:39 crc kubenswrapper[4888]: I1124 00:50:39.305257 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-775b95987-88pxg" Nov 24 00:50:39 crc kubenswrapper[4888]: I1124 00:50:39.364591 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:50:39 crc kubenswrapper[4888]: I1124 00:50:39.521726 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7d4c46565b-l2868" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerName="heat-cfnapi" containerID="cri-o://be619cd7214e54b2bf26e2c69c27ef5c7a16d950bedf4b4d10469236963ae4e8" gracePeriod=60 Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.550954 4888 generic.go:334] "Generic (PLEG): container finished" podID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerID="e3b1c60baf33e1c69193901f5b70d3bd167c112a9db60773ac0f39f047c8f838" exitCode=0 Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.551041 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69d5b6b95b-jhbt6" event={"ID":"9f2c7353-a414-429f-a6a6-f42ed6120ccf","Type":"ContainerDied","Data":"e3b1c60baf33e1c69193901f5b70d3bd167c112a9db60773ac0f39f047c8f838"} Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.838389 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-w7s5d"] Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.850308 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-w7s5d"] Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.943280 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-2btp2"] Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.944990 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.948763 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.948914 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.948996 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4kxs\" (UniqueName: \"kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.949090 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.949217 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:42 crc kubenswrapper[4888]: I1124 00:50:42.969870 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2btp2"] Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.050788 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.050884 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4kxs\" (UniqueName: \"kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.050948 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.051012 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.059621 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.061056 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.062424 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.077151 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4kxs\" (UniqueName: \"kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs\") pod \"aodh-db-sync-2btp2\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.265670 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.556572 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7d4c46565b-l2868" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.203:8000/healthcheck\": dial tcp 10.217.0.203:8000: connect: connection refused" Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.564456 4888 generic.go:334] "Generic (PLEG): container finished" podID="021b7361-39f2-47fa-b028-b1656b5fe94c" containerID="eea65ea1d457f5113f5c553d1b73980bfc751f5f5d703990dbb2ec5aeef9c994" exitCode=0 Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.564545 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"021b7361-39f2-47fa-b028-b1656b5fe94c","Type":"ContainerDied","Data":"eea65ea1d457f5113f5c553d1b73980bfc751f5f5d703990dbb2ec5aeef9c994"} Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.568118 4888 generic.go:334] "Generic (PLEG): container finished" podID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerID="be619cd7214e54b2bf26e2c69c27ef5c7a16d950bedf4b4d10469236963ae4e8" exitCode=0 Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.568211 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4c46565b-l2868" event={"ID":"6f9f41a6-bbce-4876-b1f5-b955b1c19786","Type":"ContainerDied","Data":"be619cd7214e54b2bf26e2c69c27ef5c7a16d950bedf4b4d10469236963ae4e8"} Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.571411 4888 generic.go:334] "Generic (PLEG): container finished" podID="f987cf1a-247e-411e-9d03-4a5f44a98d14" containerID="a8207573f74db48f974439d2e7b567d467fff83424b51d42ad0cdfc3ec5618ce" exitCode=0 Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.571450 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f987cf1a-247e-411e-9d03-4a5f44a98d14","Type":"ContainerDied","Data":"a8207573f74db48f974439d2e7b567d467fff83424b51d42ad0cdfc3ec5618ce"} Nov 24 00:50:43 crc kubenswrapper[4888]: I1124 00:50:43.573418 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-69d5b6b95b-jhbt6" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.204:8004/healthcheck\": dial tcp 10.217.0.204:8004: connect: connection refused" Nov 24 00:50:44 crc kubenswrapper[4888]: E1124 00:50:44.191914 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:44 crc kubenswrapper[4888]: E1124 00:50:44.194473 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:44 crc kubenswrapper[4888]: E1124 00:50:44.196273 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:44 crc kubenswrapper[4888]: E1124 00:50:44.196346 4888 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-595c95d7db-65z8t" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" Nov 24 00:50:44 crc kubenswrapper[4888]: I1124 00:50:44.263100 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166b8865-2142-4a68-852b-ac7511643e32" path="/var/lib/kubelet/pods/166b8865-2142-4a68-852b-ac7511643e32/volumes" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.769569 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.842996 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.843159 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkd4d\" (UniqueName: \"kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.843189 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.843225 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.843344 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.843378 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs\") pod \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\" (UID: \"6f9f41a6-bbce-4876-b1f5-b955b1c19786\") " Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.850236 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.853990 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d" (OuterVolumeSpecName: "kube-api-access-mkd4d") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "kube-api-access-mkd4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.940399 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.946319 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkd4d\" (UniqueName: \"kubernetes.io/projected/6f9f41a6-bbce-4876-b1f5-b955b1c19786-kube-api-access-mkd4d\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.946347 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:46 crc kubenswrapper[4888]: I1124 00:50:46.952255 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047332 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047490 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkkpr\" (UniqueName: \"kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047517 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047582 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047764 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.047841 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom\") pod \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\" (UID: \"9f2c7353-a414-429f-a6a6-f42ed6120ccf\") " Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.048240 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.054160 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.054411 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr" (OuterVolumeSpecName: "kube-api-access-qkkpr") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "kube-api-access-qkkpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.107836 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.114822 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.115328 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data" (OuterVolumeSpecName: "config-data") pod "6f9f41a6-bbce-4876-b1f5-b955b1c19786" (UID: "6f9f41a6-bbce-4876-b1f5-b955b1c19786"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.120834 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2btp2"] Nov 24 00:50:47 crc kubenswrapper[4888]: W1124 00:50:47.122944 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod631c3444_214f_4fd7_9bb9_6584c5018e82.slice/crio-b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87 WatchSource:0}: Error finding container b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87: Status 404 returned error can't find the container with id b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87 Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.142612 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.146999 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149605 4888 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149634 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149645 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f9f41a6-bbce-4876-b1f5-b955b1c19786-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149654 4888 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149665 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149674 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkkpr\" (UniqueName: \"kubernetes.io/projected/9f2c7353-a414-429f-a6a6-f42ed6120ccf-kube-api-access-qkkpr\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.149684 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.150828 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data" (OuterVolumeSpecName: "config-data") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.155401 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9f2c7353-a414-429f-a6a6-f42ed6120ccf" (UID: "9f2c7353-a414-429f-a6a6-f42ed6120ccf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.251381 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.251563 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f2c7353-a414-429f-a6a6-f42ed6120ccf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.637676 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" event={"ID":"64895ae8-ff60-4598-b404-cdd6c14c0c02","Type":"ContainerStarted","Data":"0e20725fc483a9c7ea2bde0a0c5d5c4941ea60ed74c5ee28194d1d6dc528995d"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.641886 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4c46565b-l2868" event={"ID":"6f9f41a6-bbce-4876-b1f5-b955b1c19786","Type":"ContainerDied","Data":"85f9403ee321b09862f2ea58a09d13dcf8e47a7d0d3dac754bcb459b2936a8a7"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.641946 4888 scope.go:117] "RemoveContainer" containerID="be619cd7214e54b2bf26e2c69c27ef5c7a16d950bedf4b4d10469236963ae4e8" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.642079 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4c46565b-l2868" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.661042 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f987cf1a-247e-411e-9d03-4a5f44a98d14","Type":"ContainerStarted","Data":"99ec157df64fa886827d6d886783a895ae9f77a51820cc6713a26563c88109f7"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.662086 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.685137 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" podStartSLOduration=2.982908115 podStartE2EDuration="14.685117266s" podCreationTimestamp="2025-11-24 00:50:33 +0000 UTC" firstStartedPulling="2025-11-24 00:50:34.860173418 +0000 UTC m=+1537.442857472" lastFinishedPulling="2025-11-24 00:50:46.562382579 +0000 UTC m=+1549.145066623" observedRunningTime="2025-11-24 00:50:47.669224059 +0000 UTC m=+1550.251908103" watchObservedRunningTime="2025-11-24 00:50:47.685117266 +0000 UTC m=+1550.267801310" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.690995 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"021b7361-39f2-47fa-b028-b1656b5fe94c","Type":"ContainerStarted","Data":"5868a23d9a41b11eca688597eae2dc6aaa3f7b6607e49b154a2ecdb1db9885f7"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.692603 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.703322 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69d5b6b95b-jhbt6" event={"ID":"9f2c7353-a414-429f-a6a6-f42ed6120ccf","Type":"ContainerDied","Data":"62fae4d3a1d25684384949176eec3a53f42fcd0883426470dd1e95a747c349e0"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.703516 4888 scope.go:117] "RemoveContainer" containerID="e3b1c60baf33e1c69193901f5b70d3bd167c112a9db60773ac0f39f047c8f838" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.703858 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69d5b6b95b-jhbt6" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.706761 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2btp2" event={"ID":"631c3444-214f-4fd7-9bb9-6584c5018e82","Type":"ContainerStarted","Data":"b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87"} Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.762295 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.762272537 podStartE2EDuration="41.762272537s" podCreationTimestamp="2025-11-24 00:50:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:50:47.691034573 +0000 UTC m=+1550.273718627" watchObservedRunningTime="2025-11-24 00:50:47.762272537 +0000 UTC m=+1550.344956581" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.796913 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.806641 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7d4c46565b-l2868"] Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.812214 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.812195741 podStartE2EDuration="41.812195741s" podCreationTimestamp="2025-11-24 00:50:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:50:47.743240421 +0000 UTC m=+1550.325924475" watchObservedRunningTime="2025-11-24 00:50:47.812195741 +0000 UTC m=+1550.394879785" Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.829178 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:50:47 crc kubenswrapper[4888]: I1124 00:50:47.840925 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69d5b6b95b-jhbt6"] Nov 24 00:50:48 crc kubenswrapper[4888]: I1124 00:50:48.261625 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" path="/var/lib/kubelet/pods/6f9f41a6-bbce-4876-b1f5-b955b1c19786/volumes" Nov 24 00:50:48 crc kubenswrapper[4888]: I1124 00:50:48.267581 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" path="/var/lib/kubelet/pods/9f2c7353-a414-429f-a6a6-f42ed6120ccf/volumes" Nov 24 00:50:50 crc kubenswrapper[4888]: I1124 00:50:50.245512 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:50:50 crc kubenswrapper[4888]: E1124 00:50:50.245790 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:50:51 crc kubenswrapper[4888]: I1124 00:50:51.188665 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 00:50:52 crc kubenswrapper[4888]: I1124 00:50:52.790533 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2btp2" event={"ID":"631c3444-214f-4fd7-9bb9-6584c5018e82","Type":"ContainerStarted","Data":"109eaa690a489aed893a747b15ff79dabe2790e0055a5a5f4d43e78931550e51"} Nov 24 00:50:52 crc kubenswrapper[4888]: I1124 00:50:52.827165 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-2btp2" podStartSLOduration=5.720176053 podStartE2EDuration="10.827108761s" podCreationTimestamp="2025-11-24 00:50:42 +0000 UTC" firstStartedPulling="2025-11-24 00:50:47.126342596 +0000 UTC m=+1549.709026640" lastFinishedPulling="2025-11-24 00:50:52.233275304 +0000 UTC m=+1554.815959348" observedRunningTime="2025-11-24 00:50:52.814414933 +0000 UTC m=+1555.397098977" watchObservedRunningTime="2025-11-24 00:50:52.827108761 +0000 UTC m=+1555.409792795" Nov 24 00:50:54 crc kubenswrapper[4888]: E1124 00:50:54.190860 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:54 crc kubenswrapper[4888]: E1124 00:50:54.192722 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:54 crc kubenswrapper[4888]: E1124 00:50:54.193936 4888 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 24 00:50:54 crc kubenswrapper[4888]: E1124 00:50:54.194004 4888 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-595c95d7db-65z8t" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" Nov 24 00:50:55 crc kubenswrapper[4888]: I1124 00:50:55.826620 4888 generic.go:334] "Generic (PLEG): container finished" podID="631c3444-214f-4fd7-9bb9-6584c5018e82" containerID="109eaa690a489aed893a747b15ff79dabe2790e0055a5a5f4d43e78931550e51" exitCode=0 Nov 24 00:50:55 crc kubenswrapper[4888]: I1124 00:50:55.826702 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2btp2" event={"ID":"631c3444-214f-4fd7-9bb9-6584c5018e82","Type":"ContainerDied","Data":"109eaa690a489aed893a747b15ff79dabe2790e0055a5a5f4d43e78931550e51"} Nov 24 00:50:56 crc kubenswrapper[4888]: I1124 00:50:56.564185 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 00:50:56 crc kubenswrapper[4888]: I1124 00:50:56.582981 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.276401 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.384689 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data\") pod \"631c3444-214f-4fd7-9bb9-6584c5018e82\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.384769 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4kxs\" (UniqueName: \"kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs\") pod \"631c3444-214f-4fd7-9bb9-6584c5018e82\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.385111 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts\") pod \"631c3444-214f-4fd7-9bb9-6584c5018e82\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.385189 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle\") pod \"631c3444-214f-4fd7-9bb9-6584c5018e82\" (UID: \"631c3444-214f-4fd7-9bb9-6584c5018e82\") " Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.396874 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs" (OuterVolumeSpecName: "kube-api-access-j4kxs") pod "631c3444-214f-4fd7-9bb9-6584c5018e82" (UID: "631c3444-214f-4fd7-9bb9-6584c5018e82"). InnerVolumeSpecName "kube-api-access-j4kxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.398413 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts" (OuterVolumeSpecName: "scripts") pod "631c3444-214f-4fd7-9bb9-6584c5018e82" (UID: "631c3444-214f-4fd7-9bb9-6584c5018e82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.426555 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data" (OuterVolumeSpecName: "config-data") pod "631c3444-214f-4fd7-9bb9-6584c5018e82" (UID: "631c3444-214f-4fd7-9bb9-6584c5018e82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.427881 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "631c3444-214f-4fd7-9bb9-6584c5018e82" (UID: "631c3444-214f-4fd7-9bb9-6584c5018e82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.487700 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.487739 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.487753 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631c3444-214f-4fd7-9bb9-6584c5018e82-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.487764 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4kxs\" (UniqueName: \"kubernetes.io/projected/631c3444-214f-4fd7-9bb9-6584c5018e82-kube-api-access-j4kxs\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.846958 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2btp2" event={"ID":"631c3444-214f-4fd7-9bb9-6584c5018e82","Type":"ContainerDied","Data":"b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87"} Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.846998 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b685b85fbc1b722f282d606276d2f4bbba51ddffd80accd7623a82e9c30abe87" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.847049 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2btp2" Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.850901 4888 generic.go:334] "Generic (PLEG): container finished" podID="64895ae8-ff60-4598-b404-cdd6c14c0c02" containerID="0e20725fc483a9c7ea2bde0a0c5d5c4941ea60ed74c5ee28194d1d6dc528995d" exitCode=0 Nov 24 00:50:57 crc kubenswrapper[4888]: I1124 00:50:57.850935 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" event={"ID":"64895ae8-ff60-4598-b404-cdd6c14c0c02","Type":"ContainerDied","Data":"0e20725fc483a9c7ea2bde0a0c5d5c4941ea60ed74c5ee28194d1d6dc528995d"} Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.090316 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.090566 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-api" containerID="cri-o://ee285cde682fd3ee2d78c54cefcd8eb98634ecd91ce822877265e594545fe108" gracePeriod=30 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.090610 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-listener" containerID="cri-o://9f1d9abab17e134ac69708cd99eaa668c4bdd467af737ca28b2b5125b9b4bd32" gracePeriod=30 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.090691 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-evaluator" containerID="cri-o://8de4099978c784e388027c1d253ff368974310ef4fac664c5a8259064cd39a5a" gracePeriod=30 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.090629 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-notifier" containerID="cri-o://ec323d623cb347bd09a2b25e20ca11b410d5f9d79a7685e048c188ef0c15c667" gracePeriod=30 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.862846 4888 generic.go:334] "Generic (PLEG): container finished" podID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerID="8de4099978c784e388027c1d253ff368974310ef4fac664c5a8259064cd39a5a" exitCode=0 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.863174 4888 generic.go:334] "Generic (PLEG): container finished" podID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerID="ee285cde682fd3ee2d78c54cefcd8eb98634ecd91ce822877265e594545fe108" exitCode=0 Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.862959 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerDied","Data":"8de4099978c784e388027c1d253ff368974310ef4fac664c5a8259064cd39a5a"} Nov 24 00:50:58 crc kubenswrapper[4888]: I1124 00:50:58.863239 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerDied","Data":"ee285cde682fd3ee2d78c54cefcd8eb98634ecd91ce822877265e594545fe108"} Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.342200 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.376230 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.431319 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key\") pod \"64895ae8-ff60-4598-b404-cdd6c14c0c02\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.431500 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd42s\" (UniqueName: \"kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s\") pod \"64895ae8-ff60-4598-b404-cdd6c14c0c02\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.431562 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle\") pod \"64895ae8-ff60-4598-b404-cdd6c14c0c02\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.431669 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory\") pod \"64895ae8-ff60-4598-b404-cdd6c14c0c02\" (UID: \"64895ae8-ff60-4598-b404-cdd6c14c0c02\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.438143 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "64895ae8-ff60-4598-b404-cdd6c14c0c02" (UID: "64895ae8-ff60-4598-b404-cdd6c14c0c02"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.440289 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s" (OuterVolumeSpecName: "kube-api-access-bd42s") pod "64895ae8-ff60-4598-b404-cdd6c14c0c02" (UID: "64895ae8-ff60-4598-b404-cdd6c14c0c02"). InnerVolumeSpecName "kube-api-access-bd42s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.461533 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "64895ae8-ff60-4598-b404-cdd6c14c0c02" (UID: "64895ae8-ff60-4598-b404-cdd6c14c0c02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.469575 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory" (OuterVolumeSpecName: "inventory") pod "64895ae8-ff60-4598-b404-cdd6c14c0c02" (UID: "64895ae8-ff60-4598-b404-cdd6c14c0c02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.533183 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb742\" (UniqueName: \"kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742\") pod \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.533243 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data\") pod \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.533281 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom\") pod \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.533398 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle\") pod \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\" (UID: \"6865e42c-53d9-44fa-8fb2-77e9201fc53f\") " Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.534137 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.534158 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.534171 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd42s\" (UniqueName: \"kubernetes.io/projected/64895ae8-ff60-4598-b404-cdd6c14c0c02-kube-api-access-bd42s\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.534186 4888 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64895ae8-ff60-4598-b404-cdd6c14c0c02-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.536924 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6865e42c-53d9-44fa-8fb2-77e9201fc53f" (UID: "6865e42c-53d9-44fa-8fb2-77e9201fc53f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.536977 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742" (OuterVolumeSpecName: "kube-api-access-xb742") pod "6865e42c-53d9-44fa-8fb2-77e9201fc53f" (UID: "6865e42c-53d9-44fa-8fb2-77e9201fc53f"). InnerVolumeSpecName "kube-api-access-xb742". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.570203 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6865e42c-53d9-44fa-8fb2-77e9201fc53f" (UID: "6865e42c-53d9-44fa-8fb2-77e9201fc53f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.590777 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data" (OuterVolumeSpecName: "config-data") pod "6865e42c-53d9-44fa-8fb2-77e9201fc53f" (UID: "6865e42c-53d9-44fa-8fb2-77e9201fc53f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.636536 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb742\" (UniqueName: \"kubernetes.io/projected/6865e42c-53d9-44fa-8fb2-77e9201fc53f-kube-api-access-xb742\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.636568 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.636578 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.636588 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6865e42c-53d9-44fa-8fb2-77e9201fc53f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.881369 4888 generic.go:334] "Generic (PLEG): container finished" podID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" exitCode=0 Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.881476 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-595c95d7db-65z8t" event={"ID":"6865e42c-53d9-44fa-8fb2-77e9201fc53f","Type":"ContainerDied","Data":"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634"} Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.881567 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-595c95d7db-65z8t" event={"ID":"6865e42c-53d9-44fa-8fb2-77e9201fc53f","Type":"ContainerDied","Data":"36aa6b627a9e43e8f76d68913988e887e635f5c4dae335877d044e5f5aed7d9b"} Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.881509 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-595c95d7db-65z8t" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.881608 4888 scope.go:117] "RemoveContainer" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.883890 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" event={"ID":"64895ae8-ff60-4598-b404-cdd6c14c0c02","Type":"ContainerDied","Data":"5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad"} Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.884006 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a9f0d3588ed52443673ec85f57abfe9f34e7a30dea25d88443e6a5fc71ffcad" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.883981 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.922316 4888 scope.go:117] "RemoveContainer" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.923193 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634\": container with ID starting with cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634 not found: ID does not exist" containerID="cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.923263 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634"} err="failed to get container status \"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634\": rpc error: code = NotFound desc = could not find container \"cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634\": container with ID starting with cba53ed661a517089b09647dc61fe344b833a089d4c765ee4efd7adb7f2e2634 not found: ID does not exist" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.942333 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.952495 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-595c95d7db-65z8t"] Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.968897 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2"] Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.969481 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerName="heat-cfnapi" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969508 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerName="heat-cfnapi" Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.969543 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64895ae8-ff60-4598-b404-cdd6c14c0c02" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969554 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="64895ae8-ff60-4598-b404-cdd6c14c0c02" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.969564 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631c3444-214f-4fd7-9bb9-6584c5018e82" containerName="aodh-db-sync" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969571 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="631c3444-214f-4fd7-9bb9-6584c5018e82" containerName="aodh-db-sync" Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.969583 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerName="heat-api" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969589 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerName="heat-api" Nov 24 00:50:59 crc kubenswrapper[4888]: E1124 00:50:59.969607 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969615 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969867 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2c7353-a414-429f-a6a6-f42ed6120ccf" containerName="heat-api" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969892 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="64895ae8-ff60-4598-b404-cdd6c14c0c02" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969921 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="631c3444-214f-4fd7-9bb9-6584c5018e82" containerName="aodh-db-sync" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969935 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f9f41a6-bbce-4876-b1f5-b955b1c19786" containerName="heat-cfnapi" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.969950 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" containerName="heat-engine" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.970936 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.973917 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.974119 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.974262 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.974437 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:50:59 crc kubenswrapper[4888]: I1124 00:50:59.983396 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2"] Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.146955 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.147388 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.147439 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpb6\" (UniqueName: \"kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.147475 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.249420 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.249661 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpb6\" (UniqueName: \"kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.249779 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.249908 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.253115 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.258327 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.260634 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6865e42c-53d9-44fa-8fb2-77e9201fc53f" path="/var/lib/kubelet/pods/6865e42c-53d9-44fa-8fb2-77e9201fc53f/volumes" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.267617 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.269691 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpb6\" (UniqueName: \"kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.295620 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:51:00 crc kubenswrapper[4888]: I1124 00:51:00.897230 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2"] Nov 24 00:51:01 crc kubenswrapper[4888]: I1124 00:51:01.914356 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" event={"ID":"f5e9a324-c15a-4b2e-8d60-4651313aca06","Type":"ContainerStarted","Data":"1c4855ea7bb8c5e2544a960adaa722e37260afe655c073d0f2be83032156fa78"} Nov 24 00:51:01 crc kubenswrapper[4888]: I1124 00:51:01.915127 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" event={"ID":"f5e9a324-c15a-4b2e-8d60-4651313aca06","Type":"ContainerStarted","Data":"5f081bab36c61fcda88663391850fd5b05248a21d72caa7ebbe59c57f2b72b5a"} Nov 24 00:51:01 crc kubenswrapper[4888]: I1124 00:51:01.937834 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" podStartSLOduration=2.491724871 podStartE2EDuration="2.937782405s" podCreationTimestamp="2025-11-24 00:50:59 +0000 UTC" firstStartedPulling="2025-11-24 00:51:00.902037985 +0000 UTC m=+1563.484722029" lastFinishedPulling="2025-11-24 00:51:01.348095499 +0000 UTC m=+1563.930779563" observedRunningTime="2025-11-24 00:51:01.934440691 +0000 UTC m=+1564.517124745" watchObservedRunningTime="2025-11-24 00:51:01.937782405 +0000 UTC m=+1564.520466479" Nov 24 00:51:02 crc kubenswrapper[4888]: I1124 00:51:02.930559 4888 generic.go:334] "Generic (PLEG): container finished" podID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerID="9f1d9abab17e134ac69708cd99eaa668c4bdd467af737ca28b2b5125b9b4bd32" exitCode=0 Nov 24 00:51:02 crc kubenswrapper[4888]: I1124 00:51:02.930939 4888 generic.go:334] "Generic (PLEG): container finished" podID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerID="ec323d623cb347bd09a2b25e20ca11b410d5f9d79a7685e048c188ef0c15c667" exitCode=0 Nov 24 00:51:02 crc kubenswrapper[4888]: I1124 00:51:02.930752 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerDied","Data":"9f1d9abab17e134ac69708cd99eaa668c4bdd467af737ca28b2b5125b9b4bd32"} Nov 24 00:51:02 crc kubenswrapper[4888]: I1124 00:51:02.931049 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerDied","Data":"ec323d623cb347bd09a2b25e20ca11b410d5f9d79a7685e048c188ef0c15c667"} Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.038504 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112098 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112222 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112319 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112371 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112419 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.112449 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ndn7\" (UniqueName: \"kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7\") pod \"5645bf92-6527-49a8-90e7-721ae2eaedf6\" (UID: \"5645bf92-6527-49a8-90e7-721ae2eaedf6\") " Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.121509 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7" (OuterVolumeSpecName: "kube-api-access-4ndn7") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "kube-api-access-4ndn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.142489 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts" (OuterVolumeSpecName: "scripts") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.186406 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.202150 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.215983 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.216012 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.216023 4888 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.216031 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ndn7\" (UniqueName: \"kubernetes.io/projected/5645bf92-6527-49a8-90e7-721ae2eaedf6-kube-api-access-4ndn7\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.253020 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data" (OuterVolumeSpecName: "config-data") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.258764 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5645bf92-6527-49a8-90e7-721ae2eaedf6" (UID: "5645bf92-6527-49a8-90e7-721ae2eaedf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.318479 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.318510 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5645bf92-6527-49a8-90e7-721ae2eaedf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.942296 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5645bf92-6527-49a8-90e7-721ae2eaedf6","Type":"ContainerDied","Data":"2e9fa882b485b4cb5309fe0a2a2d2db25fed6080c8ce767b94dde09a4f33e1a9"} Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.942353 4888 scope.go:117] "RemoveContainer" containerID="9f1d9abab17e134ac69708cd99eaa668c4bdd467af737ca28b2b5125b9b4bd32" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.942514 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.982739 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.993203 4888 scope.go:117] "RemoveContainer" containerID="ec323d623cb347bd09a2b25e20ca11b410d5f9d79a7685e048c188ef0c15c667" Nov 24 00:51:03 crc kubenswrapper[4888]: I1124 00:51:03.999161 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.026178 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 00:51:04 crc kubenswrapper[4888]: E1124 00:51:04.026755 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-api" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.026780 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-api" Nov 24 00:51:04 crc kubenswrapper[4888]: E1124 00:51:04.026833 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-listener" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.026843 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-listener" Nov 24 00:51:04 crc kubenswrapper[4888]: E1124 00:51:04.026863 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-notifier" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.026871 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-notifier" Nov 24 00:51:04 crc kubenswrapper[4888]: E1124 00:51:04.026892 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-evaluator" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.026900 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-evaluator" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.027135 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-listener" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.027161 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-evaluator" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.027179 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-notifier" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.031001 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" containerName="aodh-api" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.033334 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.033444 4888 scope.go:117] "RemoveContainer" containerID="8de4099978c784e388027c1d253ff368974310ef4fac664c5a8259064cd39a5a" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.035725 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.036029 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.036181 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.036662 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-t9vsv" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.038940 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.057397 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.080583 4888 scope.go:117] "RemoveContainer" containerID="ee285cde682fd3ee2d78c54cefcd8eb98634ecd91ce822877265e594545fe108" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.134894 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-internal-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.134992 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.135022 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f94md\" (UniqueName: \"kubernetes.io/projected/24be4c10-1e0a-4452-b76b-1d979796231d-kube-api-access-f94md\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.135058 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-config-data\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.135073 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-public-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.135163 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-scripts\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238248 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-internal-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238361 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f94md\" (UniqueName: \"kubernetes.io/projected/24be4c10-1e0a-4452-b76b-1d979796231d-kube-api-access-f94md\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238436 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-config-data\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238457 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-public-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.238504 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-scripts\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.242659 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-scripts\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.243253 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.245539 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-public-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.245739 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:51:04 crc kubenswrapper[4888]: E1124 00:51:04.246049 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.247602 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-internal-tls-certs\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.248837 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24be4c10-1e0a-4452-b76b-1d979796231d-config-data\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.255930 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f94md\" (UniqueName: \"kubernetes.io/projected/24be4c10-1e0a-4452-b76b-1d979796231d-kube-api-access-f94md\") pod \"aodh-0\" (UID: \"24be4c10-1e0a-4452-b76b-1d979796231d\") " pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.259215 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5645bf92-6527-49a8-90e7-721ae2eaedf6" path="/var/lib/kubelet/pods/5645bf92-6527-49a8-90e7-721ae2eaedf6/volumes" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.359830 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.819556 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 00:51:04 crc kubenswrapper[4888]: I1124 00:51:04.957295 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"24be4c10-1e0a-4452-b76b-1d979796231d","Type":"ContainerStarted","Data":"70a2de6d2559387440bfb074ae9dd41f22496d0fc6b3790a36f4227f7ba711bb"} Nov 24 00:51:07 crc kubenswrapper[4888]: I1124 00:51:07.991519 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"24be4c10-1e0a-4452-b76b-1d979796231d","Type":"ContainerStarted","Data":"e24edfcdec24105d520720eb5fe0767e8e01fcce36ff89a8daf7048a7d2e18ab"} Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.248127 4888 scope.go:117] "RemoveContainer" containerID="ffa92dbfd86603ffbe9031bbd1601e583f01c66d3d29826cb69f3b20bbb67ec4" Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.316162 4888 scope.go:117] "RemoveContainer" containerID="e6b862ec2e5da8da51e27851824a109f095739d5a1d306e94cddc7a48a7fdcac" Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.348325 4888 scope.go:117] "RemoveContainer" containerID="34f418c2d731b62722a2089e7af7bede3064c9909b7e01445dfcb5b698fccf88" Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.393668 4888 scope.go:117] "RemoveContainer" containerID="0a660a955587e6605222ffbc6accc58c21bc2de4195fe4909a40f8bb2e9fef58" Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.573092 4888 scope.go:117] "RemoveContainer" containerID="f00812e0f8369a19c2f0bf3b384086717a2595fd39171ef4559011cc873d78e6" Nov 24 00:51:08 crc kubenswrapper[4888]: I1124 00:51:08.627278 4888 scope.go:117] "RemoveContainer" containerID="af2948780bd3f2b73a8921acdf3979ecc822211ca95d831eb7678f81d02a4df9" Nov 24 00:51:09 crc kubenswrapper[4888]: I1124 00:51:09.032399 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"24be4c10-1e0a-4452-b76b-1d979796231d","Type":"ContainerStarted","Data":"3917a6b68f31e372ebd3fb88be6be60bb6e7c098b4164429fd849098e0ee5602"} Nov 24 00:51:10 crc kubenswrapper[4888]: I1124 00:51:10.057854 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"24be4c10-1e0a-4452-b76b-1d979796231d","Type":"ContainerStarted","Data":"cd009c33d35a80600e292aaebf9edbd38fdbf8af365816c2cbf11b2ebbed88f8"} Nov 24 00:51:12 crc kubenswrapper[4888]: I1124 00:51:12.078301 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"24be4c10-1e0a-4452-b76b-1d979796231d","Type":"ContainerStarted","Data":"ab7e0fc74746289f35232aea541a8bd253a416ac475d7482d889211ed6d294b8"} Nov 24 00:51:12 crc kubenswrapper[4888]: I1124 00:51:12.141248 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.130583057 podStartE2EDuration="9.141218128s" podCreationTimestamp="2025-11-24 00:51:03 +0000 UTC" firstStartedPulling="2025-11-24 00:51:04.817867939 +0000 UTC m=+1567.400551983" lastFinishedPulling="2025-11-24 00:51:10.82850301 +0000 UTC m=+1573.411187054" observedRunningTime="2025-11-24 00:51:12.111275987 +0000 UTC m=+1574.693960041" watchObservedRunningTime="2025-11-24 00:51:12.141218128 +0000 UTC m=+1574.723902172" Nov 24 00:51:15 crc kubenswrapper[4888]: I1124 00:51:15.245658 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:51:15 crc kubenswrapper[4888]: E1124 00:51:15.247913 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:51:26 crc kubenswrapper[4888]: I1124 00:51:26.245960 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:51:26 crc kubenswrapper[4888]: E1124 00:51:26.247178 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:51:39 crc kubenswrapper[4888]: I1124 00:51:39.245473 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:51:39 crc kubenswrapper[4888]: E1124 00:51:39.247932 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:51:52 crc kubenswrapper[4888]: I1124 00:51:52.246102 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:51:52 crc kubenswrapper[4888]: E1124 00:51:52.247177 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:52:04 crc kubenswrapper[4888]: I1124 00:52:04.246211 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:52:04 crc kubenswrapper[4888]: E1124 00:52:04.247237 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:52:08 crc kubenswrapper[4888]: I1124 00:52:08.927802 4888 scope.go:117] "RemoveContainer" containerID="7e63107fa917588d5227a8528b69ee5c1e2683db8651aaa589458a02d317ac88" Nov 24 00:52:08 crc kubenswrapper[4888]: I1124 00:52:08.967943 4888 scope.go:117] "RemoveContainer" containerID="a2f94b083b607c53a4955f05771ed86ec1873c0f906a2a4818209a87a66e83f1" Nov 24 00:52:09 crc kubenswrapper[4888]: I1124 00:52:09.061146 4888 scope.go:117] "RemoveContainer" containerID="637fe1043e79450f555a42fea83447bf5def0711407e00f87b4463a927affed5" Nov 24 00:52:19 crc kubenswrapper[4888]: I1124 00:52:19.244782 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:52:19 crc kubenswrapper[4888]: E1124 00:52:19.245805 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:52:34 crc kubenswrapper[4888]: I1124 00:52:34.246439 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:52:34 crc kubenswrapper[4888]: E1124 00:52:34.249458 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:52:45 crc kubenswrapper[4888]: I1124 00:52:45.246430 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:52:45 crc kubenswrapper[4888]: E1124 00:52:45.247213 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:53:00 crc kubenswrapper[4888]: I1124 00:53:00.246708 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:53:00 crc kubenswrapper[4888]: E1124 00:53:00.247768 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:53:13 crc kubenswrapper[4888]: I1124 00:53:13.245744 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:53:13 crc kubenswrapper[4888]: E1124 00:53:13.247186 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:53:25 crc kubenswrapper[4888]: I1124 00:53:25.245389 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:53:25 crc kubenswrapper[4888]: E1124 00:53:25.246532 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:53:40 crc kubenswrapper[4888]: I1124 00:53:40.245438 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:53:40 crc kubenswrapper[4888]: E1124 00:53:40.246677 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:53:51 crc kubenswrapper[4888]: I1124 00:53:51.245338 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:53:51 crc kubenswrapper[4888]: E1124 00:53:51.246304 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:54:04 crc kubenswrapper[4888]: I1124 00:54:04.292068 4888 generic.go:334] "Generic (PLEG): container finished" podID="f5e9a324-c15a-4b2e-8d60-4651313aca06" containerID="1c4855ea7bb8c5e2544a960adaa722e37260afe655c073d0f2be83032156fa78" exitCode=0 Nov 24 00:54:04 crc kubenswrapper[4888]: I1124 00:54:04.292194 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" event={"ID":"f5e9a324-c15a-4b2e-8d60-4651313aca06","Type":"ContainerDied","Data":"1c4855ea7bb8c5e2544a960adaa722e37260afe655c073d0f2be83032156fa78"} Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.245967 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:54:05 crc kubenswrapper[4888]: E1124 00:54:05.246437 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.840680 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.961510 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory\") pod \"f5e9a324-c15a-4b2e-8d60-4651313aca06\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.961704 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxpb6\" (UniqueName: \"kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6\") pod \"f5e9a324-c15a-4b2e-8d60-4651313aca06\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.961727 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle\") pod \"f5e9a324-c15a-4b2e-8d60-4651313aca06\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.961798 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key\") pod \"f5e9a324-c15a-4b2e-8d60-4651313aca06\" (UID: \"f5e9a324-c15a-4b2e-8d60-4651313aca06\") " Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.968300 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f5e9a324-c15a-4b2e-8d60-4651313aca06" (UID: "f5e9a324-c15a-4b2e-8d60-4651313aca06"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:54:05 crc kubenswrapper[4888]: I1124 00:54:05.969124 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6" (OuterVolumeSpecName: "kube-api-access-fxpb6") pod "f5e9a324-c15a-4b2e-8d60-4651313aca06" (UID: "f5e9a324-c15a-4b2e-8d60-4651313aca06"). InnerVolumeSpecName "kube-api-access-fxpb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.005433 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5e9a324-c15a-4b2e-8d60-4651313aca06" (UID: "f5e9a324-c15a-4b2e-8d60-4651313aca06"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.030443 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory" (OuterVolumeSpecName: "inventory") pod "f5e9a324-c15a-4b2e-8d60-4651313aca06" (UID: "f5e9a324-c15a-4b2e-8d60-4651313aca06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.063886 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.063919 4888 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.063934 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxpb6\" (UniqueName: \"kubernetes.io/projected/f5e9a324-c15a-4b2e-8d60-4651313aca06-kube-api-access-fxpb6\") on node \"crc\" DevicePath \"\"" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.063945 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5e9a324-c15a-4b2e-8d60-4651313aca06-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.338082 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" event={"ID":"f5e9a324-c15a-4b2e-8d60-4651313aca06","Type":"ContainerDied","Data":"5f081bab36c61fcda88663391850fd5b05248a21d72caa7ebbe59c57f2b72b5a"} Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.338142 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f081bab36c61fcda88663391850fd5b05248a21d72caa7ebbe59c57f2b72b5a" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.338220 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.454354 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r"] Nov 24 00:54:06 crc kubenswrapper[4888]: E1124 00:54:06.455138 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e9a324-c15a-4b2e-8d60-4651313aca06" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.455171 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e9a324-c15a-4b2e-8d60-4651313aca06" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.455517 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e9a324-c15a-4b2e-8d60-4651313aca06" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.456971 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.461571 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.462475 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.463177 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.463463 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.472444 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r"] Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.576011 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.576114 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcjhq\" (UniqueName: \"kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.576216 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.679002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.679094 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcjhq\" (UniqueName: \"kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.679205 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.685778 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.686996 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.703908 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcjhq\" (UniqueName: \"kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:06 crc kubenswrapper[4888]: I1124 00:54:06.795105 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:54:07 crc kubenswrapper[4888]: I1124 00:54:07.458221 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r"] Nov 24 00:54:07 crc kubenswrapper[4888]: W1124 00:54:07.467314 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94505667_e4ca_4420_bb6b_bff063bab9ec.slice/crio-5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16 WatchSource:0}: Error finding container 5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16: Status 404 returned error can't find the container with id 5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16 Nov 24 00:54:08 crc kubenswrapper[4888]: I1124 00:54:08.365086 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" event={"ID":"94505667-e4ca-4420-bb6b-bff063bab9ec","Type":"ContainerStarted","Data":"840c9406f811f5d10bf819048ed255dcc7b6891c3b8028c5ef40822f7c5d560f"} Nov 24 00:54:08 crc kubenswrapper[4888]: I1124 00:54:08.365353 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" event={"ID":"94505667-e4ca-4420-bb6b-bff063bab9ec","Type":"ContainerStarted","Data":"5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16"} Nov 24 00:54:08 crc kubenswrapper[4888]: I1124 00:54:08.383276 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" podStartSLOduration=1.879515769 podStartE2EDuration="2.383248633s" podCreationTimestamp="2025-11-24 00:54:06 +0000 UTC" firstStartedPulling="2025-11-24 00:54:07.470232658 +0000 UTC m=+1750.052916702" lastFinishedPulling="2025-11-24 00:54:07.973965492 +0000 UTC m=+1750.556649566" observedRunningTime="2025-11-24 00:54:08.380902337 +0000 UTC m=+1750.963586401" watchObservedRunningTime="2025-11-24 00:54:08.383248633 +0000 UTC m=+1750.965932697" Nov 24 00:54:09 crc kubenswrapper[4888]: I1124 00:54:09.198587 4888 scope.go:117] "RemoveContainer" containerID="4e1ec1f25ca75c582b1065a410999ea908110682881e906b7332c6d976e42589" Nov 24 00:54:17 crc kubenswrapper[4888]: I1124 00:54:17.087457 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jpqvm"] Nov 24 00:54:17 crc kubenswrapper[4888]: I1124 00:54:17.103097 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-aa04-account-create-c6q7c"] Nov 24 00:54:17 crc kubenswrapper[4888]: I1124 00:54:17.113636 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-aa04-account-create-c6q7c"] Nov 24 00:54:17 crc kubenswrapper[4888]: I1124 00:54:17.122074 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jpqvm"] Nov 24 00:54:18 crc kubenswrapper[4888]: I1124 00:54:18.258514 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af2ab20-ea96-42e5-9aa1-3a319b4a5630" path="/var/lib/kubelet/pods/1af2ab20-ea96-42e5-9aa1-3a319b4a5630/volumes" Nov 24 00:54:18 crc kubenswrapper[4888]: I1124 00:54:18.261584 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47fc5340-6cad-4102-8d70-c57ea387f607" path="/var/lib/kubelet/pods/47fc5340-6cad-4102-8d70-c57ea387f607/volumes" Nov 24 00:54:19 crc kubenswrapper[4888]: I1124 00:54:19.245327 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:54:19 crc kubenswrapper[4888]: E1124 00:54:19.245667 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.034371 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wx4qp"] Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.047002 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3726-account-create-7nrm5"] Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.057926 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3726-account-create-7nrm5"] Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.069052 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wx4qp"] Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.257760 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0faaac94-c8d8-4aa3-a321-250f7bf4c252" path="/var/lib/kubelet/pods/0faaac94-c8d8-4aa3-a321-250f7bf4c252/volumes" Nov 24 00:54:20 crc kubenswrapper[4888]: I1124 00:54:20.261747 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9" path="/var/lib/kubelet/pods/4e7e7b64-ea3d-411c-8a1e-d5c321fe34c9/volumes" Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.040747 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vm5vz"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.079101 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-be14-account-create-jqd96"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.088261 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-7vxql"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.097864 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0298-account-create-bt4k8"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.106548 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-7vxql"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.115009 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-be14-account-create-jqd96"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.123983 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0298-account-create-bt4k8"] Nov 24 00:54:21 crc kubenswrapper[4888]: I1124 00:54:21.131745 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vm5vz"] Nov 24 00:54:22 crc kubenswrapper[4888]: I1124 00:54:22.282251 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0051d661-e49b-4b29-8532-a975ba76ba66" path="/var/lib/kubelet/pods/0051d661-e49b-4b29-8532-a975ba76ba66/volumes" Nov 24 00:54:22 crc kubenswrapper[4888]: I1124 00:54:22.302627 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55d0fd22-3184-4fd6-8624-cbe4c1ce970c" path="/var/lib/kubelet/pods/55d0fd22-3184-4fd6-8624-cbe4c1ce970c/volumes" Nov 24 00:54:22 crc kubenswrapper[4888]: I1124 00:54:22.319409 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7525a91c-0d49-42bf-b697-3bc21bf8769f" path="/var/lib/kubelet/pods/7525a91c-0d49-42bf-b697-3bc21bf8769f/volumes" Nov 24 00:54:22 crc kubenswrapper[4888]: I1124 00:54:22.343446 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50db534-1d78-4650-ba32-f6c4eaedc101" path="/var/lib/kubelet/pods/e50db534-1d78-4650-ba32-f6c4eaedc101/volumes" Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.055105 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-a03d-account-create-2tm5v"] Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.064207 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-4s525"] Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.074837 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-4s525"] Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.082183 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-a03d-account-create-2tm5v"] Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.268159 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a98ed1-1482-4368-8d18-8ae005b81b6b" path="/var/lib/kubelet/pods/32a98ed1-1482-4368-8d18-8ae005b81b6b/volumes" Nov 24 00:54:28 crc kubenswrapper[4888]: I1124 00:54:28.273399 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e43b89-661f-439f-8996-e5f0501657b7" path="/var/lib/kubelet/pods/f9e43b89-661f-439f-8996-e5f0501657b7/volumes" Nov 24 00:54:30 crc kubenswrapper[4888]: I1124 00:54:30.247858 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:54:30 crc kubenswrapper[4888]: E1124 00:54:30.248453 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:54:42 crc kubenswrapper[4888]: I1124 00:54:42.245504 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:54:42 crc kubenswrapper[4888]: E1124 00:54:42.246351 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 00:54:52 crc kubenswrapper[4888]: I1124 00:54:52.053667 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pn29n"] Nov 24 00:54:52 crc kubenswrapper[4888]: I1124 00:54:52.063690 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pn29n"] Nov 24 00:54:52 crc kubenswrapper[4888]: I1124 00:54:52.260284 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3294b25b-9e52-43e0-ba6e-87b510640945" path="/var/lib/kubelet/pods/3294b25b-9e52-43e0-ba6e-87b510640945/volumes" Nov 24 00:54:53 crc kubenswrapper[4888]: I1124 00:54:53.029469 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-58e8-account-create-ths8q"] Nov 24 00:54:53 crc kubenswrapper[4888]: I1124 00:54:53.042710 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-58e8-account-create-ths8q"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.036176 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rl79r"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.052401 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-b919-account-create-pjsgg"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.067894 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-792e-account-create-9xf57"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.075330 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-792e-account-create-9xf57"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.085713 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-gzwq7"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.094623 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rl79r"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.103490 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-gzwq7"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.112346 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-b919-account-create-pjsgg"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.120846 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xsvt9"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.129297 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2f74-account-create-kpjmw"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.137445 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xsvt9"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.144787 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2f74-account-create-kpjmw"] Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.265506 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05103e58-6551-4742-ba88-12e3669fbee3" path="/var/lib/kubelet/pods/05103e58-6551-4742-ba88-12e3669fbee3/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.268322 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47da0c31-1bfa-4738-a4bd-fcf0e82a3e47" path="/var/lib/kubelet/pods/47da0c31-1bfa-4738-a4bd-fcf0e82a3e47/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.270950 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42" path="/var/lib/kubelet/pods/5bee8e14-3ef4-40a8-a9a9-e05d2daeeb42/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.274029 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61345e71-eb59-4bef-a890-65ed5a2e99b7" path="/var/lib/kubelet/pods/61345e71-eb59-4bef-a890-65ed5a2e99b7/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.275833 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d1470a-a999-4afe-a7e4-514914054767" path="/var/lib/kubelet/pods/83d1470a-a999-4afe-a7e4-514914054767/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.278614 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a" path="/var/lib/kubelet/pods/a53f0b5b-a751-4515-9c63-c1c3a1fdcc6a/volumes" Nov 24 00:54:54 crc kubenswrapper[4888]: I1124 00:54:54.280372 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9609549-6956-4a00-a921-5ab8f6b0511f" path="/var/lib/kubelet/pods/b9609549-6956-4a00-a921-5ab8f6b0511f/volumes" Nov 24 00:54:55 crc kubenswrapper[4888]: I1124 00:54:55.245026 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:54:55 crc kubenswrapper[4888]: I1124 00:54:55.974159 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63"} Nov 24 00:54:57 crc kubenswrapper[4888]: I1124 00:54:57.044995 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-72247"] Nov 24 00:54:57 crc kubenswrapper[4888]: I1124 00:54:57.059738 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-72247"] Nov 24 00:54:58 crc kubenswrapper[4888]: I1124 00:54:58.274305 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff49e8b-d3e9-453a-8adb-858fba7c7418" path="/var/lib/kubelet/pods/4ff49e8b-d3e9-453a-8adb-858fba7c7418/volumes" Nov 24 00:55:02 crc kubenswrapper[4888]: I1124 00:55:02.044333 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xqt9t"] Nov 24 00:55:02 crc kubenswrapper[4888]: I1124 00:55:02.056868 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xqt9t"] Nov 24 00:55:02 crc kubenswrapper[4888]: I1124 00:55:02.268418 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1cd14a9-9f24-436f-b96f-ea9457548844" path="/var/lib/kubelet/pods/c1cd14a9-9f24-436f-b96f-ea9457548844/volumes" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.332515 4888 scope.go:117] "RemoveContainer" containerID="4638da5c25e3048b04ac408c4f4b75cbe01cca5fbf3d92b2cf5cbef74054986d" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.409475 4888 scope.go:117] "RemoveContainer" containerID="e60366f2b38f8087ea2901a120a8b3d73cb01239b68f9fe7d2e68056ed70163b" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.449329 4888 scope.go:117] "RemoveContainer" containerID="65b776c2e8937ce2b76d404fc17bb108320f618ade0ccf8082b9ee1597a23294" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.489012 4888 scope.go:117] "RemoveContainer" containerID="bc40d1929d02ba606926d22c27ab8db80cea6f31bc956924dd08dbfbb2eb2919" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.569123 4888 scope.go:117] "RemoveContainer" containerID="ff0aef7e797526e500740b14215527b7679a926ee4d9715bcc2c71bac03f5e95" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.594608 4888 scope.go:117] "RemoveContainer" containerID="74f826d0db9325ba785288ad9b49740371cc443fc4c26b14aa8383326770fe40" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.645693 4888 scope.go:117] "RemoveContainer" containerID="68cd37fe99d620c36b75d67ae7670ea2f45f4aaf8b1bdab7191ff847cdd6ad35" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.669583 4888 scope.go:117] "RemoveContainer" containerID="89c2bc8e71de1d786fdf59a7f7525d85409394bcf748b023ab534a3695febbb0" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.700016 4888 scope.go:117] "RemoveContainer" containerID="53e4353af5fe051831265cc7b1f8feee569e5da881744ee8e63c4ef73f5aa418" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.728051 4888 scope.go:117] "RemoveContainer" containerID="28ae2080e28b50f3f96990a49d9ba8d389a9e74f704786ddeac56f1ceb654c6a" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.757551 4888 scope.go:117] "RemoveContainer" containerID="6f1b561bb345b91b75fea7c96d15e54444c5e6849bf646307d59120c3e07ebf2" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.790778 4888 scope.go:117] "RemoveContainer" containerID="03bc9f2bfed81b04f3a9eda68d0ecbeca10ab83fd362b2834f7c83bdff201f7d" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.817653 4888 scope.go:117] "RemoveContainer" containerID="5909d20f57ef12fa2c07fe6d1e62465c189ae01a74f7deaadd8dc46213291040" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.848973 4888 scope.go:117] "RemoveContainer" containerID="cea07138e0b390031e19f6d87e7d0a447aa791ec115d30341211198230033db1" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.872268 4888 scope.go:117] "RemoveContainer" containerID="658b6f8d0862b4c0e48c2fc7ce2229cc5c1eeab90483ccacbc617f680dd43442" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.899888 4888 scope.go:117] "RemoveContainer" containerID="acfcf3f1b38a010132c1bf6b4f04348e29b792ea91138fbe5de6d2fdcc87352a" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.931936 4888 scope.go:117] "RemoveContainer" containerID="fa2366f9cb366682b4a51a13373c728bfdd42eb67fc98c4ad14115074093fe0f" Nov 24 00:55:09 crc kubenswrapper[4888]: I1124 00:55:09.961579 4888 scope.go:117] "RemoveContainer" containerID="12069f009ddaee72158ce58e2208ca1dd5c0ed504c9d6de9b93c94c8fdefda1b" Nov 24 00:55:10 crc kubenswrapper[4888]: I1124 00:55:10.002310 4888 scope.go:117] "RemoveContainer" containerID="03708822d89411fc1b16c27682685f4c1f9a639244b751aceb5c0a145f04ea9b" Nov 24 00:55:10 crc kubenswrapper[4888]: I1124 00:55:10.029425 4888 scope.go:117] "RemoveContainer" containerID="fa8055c93cd089d42c5250a5fc961c7aeac793004c67ba37727fb561c44fc754" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.187069 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.191461 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.207382 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.261541 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.261682 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7mt4\" (UniqueName: \"kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.261743 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.363069 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.363224 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7mt4\" (UniqueName: \"kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.363342 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.363514 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.364390 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.382789 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7mt4\" (UniqueName: \"kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4\") pod \"community-operators-jstsv\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:11 crc kubenswrapper[4888]: I1124 00:55:11.525249 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:12 crc kubenswrapper[4888]: I1124 00:55:12.056833 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:12 crc kubenswrapper[4888]: W1124 00:55:12.063466 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24bf2072_516b_4fc9_bec2_2b8910f1077b.slice/crio-f61eb0ad258c3ab6ca273446a218fcafdc9d1dfe5dc59d44ee7be3f09b028494 WatchSource:0}: Error finding container f61eb0ad258c3ab6ca273446a218fcafdc9d1dfe5dc59d44ee7be3f09b028494: Status 404 returned error can't find the container with id f61eb0ad258c3ab6ca273446a218fcafdc9d1dfe5dc59d44ee7be3f09b028494 Nov 24 00:55:12 crc kubenswrapper[4888]: I1124 00:55:12.242483 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerStarted","Data":"f61eb0ad258c3ab6ca273446a218fcafdc9d1dfe5dc59d44ee7be3f09b028494"} Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.262728 4888 generic.go:334] "Generic (PLEG): container finished" podID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerID="65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa" exitCode=0 Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.262851 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerDied","Data":"65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa"} Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.268098 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.395640 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.399166 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.410672 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.509329 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh6zp\" (UniqueName: \"kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.509403 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.509503 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.612099 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh6zp\" (UniqueName: \"kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.612173 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.612285 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.612912 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.613117 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.637120 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh6zp\" (UniqueName: \"kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp\") pod \"redhat-marketplace-v4k6w\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:13 crc kubenswrapper[4888]: I1124 00:55:13.736454 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:14 crc kubenswrapper[4888]: W1124 00:55:14.257138 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc2033cd_a3c5_46ca_9a1e_67b0cec4efc4.slice/crio-e0be21c6a1689a0d784f8d12c97741418c7698949a48c45036a26917e6513277 WatchSource:0}: Error finding container e0be21c6a1689a0d784f8d12c97741418c7698949a48c45036a26917e6513277: Status 404 returned error can't find the container with id e0be21c6a1689a0d784f8d12c97741418c7698949a48c45036a26917e6513277 Nov 24 00:55:14 crc kubenswrapper[4888]: I1124 00:55:14.271383 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:14 crc kubenswrapper[4888]: I1124 00:55:14.277597 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerStarted","Data":"e0be21c6a1689a0d784f8d12c97741418c7698949a48c45036a26917e6513277"} Nov 24 00:55:14 crc kubenswrapper[4888]: I1124 00:55:14.280306 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerStarted","Data":"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5"} Nov 24 00:55:15 crc kubenswrapper[4888]: I1124 00:55:15.296899 4888 generic.go:334] "Generic (PLEG): container finished" podID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerID="7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79" exitCode=0 Nov 24 00:55:15 crc kubenswrapper[4888]: I1124 00:55:15.296999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerDied","Data":"7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79"} Nov 24 00:55:16 crc kubenswrapper[4888]: I1124 00:55:16.308791 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerStarted","Data":"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf"} Nov 24 00:55:16 crc kubenswrapper[4888]: I1124 00:55:16.311854 4888 generic.go:334] "Generic (PLEG): container finished" podID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerID="8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5" exitCode=0 Nov 24 00:55:16 crc kubenswrapper[4888]: I1124 00:55:16.311906 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerDied","Data":"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5"} Nov 24 00:55:17 crc kubenswrapper[4888]: I1124 00:55:17.326673 4888 generic.go:334] "Generic (PLEG): container finished" podID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerID="22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf" exitCode=0 Nov 24 00:55:17 crc kubenswrapper[4888]: I1124 00:55:17.326755 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerDied","Data":"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf"} Nov 24 00:55:17 crc kubenswrapper[4888]: I1124 00:55:17.331373 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerStarted","Data":"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72"} Nov 24 00:55:17 crc kubenswrapper[4888]: I1124 00:55:17.382927 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jstsv" podStartSLOduration=2.943311123 podStartE2EDuration="6.382906346s" podCreationTimestamp="2025-11-24 00:55:11 +0000 UTC" firstStartedPulling="2025-11-24 00:55:13.267746824 +0000 UTC m=+1815.850430878" lastFinishedPulling="2025-11-24 00:55:16.707342057 +0000 UTC m=+1819.290026101" observedRunningTime="2025-11-24 00:55:17.380711994 +0000 UTC m=+1819.963396108" watchObservedRunningTime="2025-11-24 00:55:17.382906346 +0000 UTC m=+1819.965590400" Nov 24 00:55:18 crc kubenswrapper[4888]: I1124 00:55:18.347798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerStarted","Data":"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558"} Nov 24 00:55:18 crc kubenswrapper[4888]: I1124 00:55:18.379505 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v4k6w" podStartSLOduration=2.784922 podStartE2EDuration="5.379484567s" podCreationTimestamp="2025-11-24 00:55:13 +0000 UTC" firstStartedPulling="2025-11-24 00:55:15.299242353 +0000 UTC m=+1817.881926407" lastFinishedPulling="2025-11-24 00:55:17.89380493 +0000 UTC m=+1820.476488974" observedRunningTime="2025-11-24 00:55:18.364758973 +0000 UTC m=+1820.947443037" watchObservedRunningTime="2025-11-24 00:55:18.379484567 +0000 UTC m=+1820.962168611" Nov 24 00:55:21 crc kubenswrapper[4888]: I1124 00:55:21.407955 4888 generic.go:334] "Generic (PLEG): container finished" podID="94505667-e4ca-4420-bb6b-bff063bab9ec" containerID="840c9406f811f5d10bf819048ed255dcc7b6891c3b8028c5ef40822f7c5d560f" exitCode=0 Nov 24 00:55:21 crc kubenswrapper[4888]: I1124 00:55:21.408068 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" event={"ID":"94505667-e4ca-4420-bb6b-bff063bab9ec","Type":"ContainerDied","Data":"840c9406f811f5d10bf819048ed255dcc7b6891c3b8028c5ef40822f7c5d560f"} Nov 24 00:55:21 crc kubenswrapper[4888]: I1124 00:55:21.525536 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:21 crc kubenswrapper[4888]: I1124 00:55:21.526619 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:22 crc kubenswrapper[4888]: I1124 00:55:22.605331 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jstsv" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="registry-server" probeResult="failure" output=< Nov 24 00:55:22 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:55:22 crc kubenswrapper[4888]: > Nov 24 00:55:22 crc kubenswrapper[4888]: I1124 00:55:22.946838 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.132122 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key\") pod \"94505667-e4ca-4420-bb6b-bff063bab9ec\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.132248 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory\") pod \"94505667-e4ca-4420-bb6b-bff063bab9ec\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.132434 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcjhq\" (UniqueName: \"kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq\") pod \"94505667-e4ca-4420-bb6b-bff063bab9ec\" (UID: \"94505667-e4ca-4420-bb6b-bff063bab9ec\") " Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.138606 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq" (OuterVolumeSpecName: "kube-api-access-hcjhq") pod "94505667-e4ca-4420-bb6b-bff063bab9ec" (UID: "94505667-e4ca-4420-bb6b-bff063bab9ec"). InnerVolumeSpecName "kube-api-access-hcjhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.166729 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory" (OuterVolumeSpecName: "inventory") pod "94505667-e4ca-4420-bb6b-bff063bab9ec" (UID: "94505667-e4ca-4420-bb6b-bff063bab9ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.172345 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94505667-e4ca-4420-bb6b-bff063bab9ec" (UID: "94505667-e4ca-4420-bb6b-bff063bab9ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.235442 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcjhq\" (UniqueName: \"kubernetes.io/projected/94505667-e4ca-4420-bb6b-bff063bab9ec-kube-api-access-hcjhq\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.235793 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.235991 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94505667-e4ca-4420-bb6b-bff063bab9ec-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.433196 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" event={"ID":"94505667-e4ca-4420-bb6b-bff063bab9ec","Type":"ContainerDied","Data":"5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16"} Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.433241 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b019a0a87d404ab168856fe188a2f5ef48ff52d74146fbfdccb162a72384f16" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.433266 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.533071 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4"] Nov 24 00:55:23 crc kubenswrapper[4888]: E1124 00:55:23.536621 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94505667-e4ca-4420-bb6b-bff063bab9ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.536645 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="94505667-e4ca-4420-bb6b-bff063bab9ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.536935 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="94505667-e4ca-4420-bb6b-bff063bab9ec" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.537702 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.540381 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.540441 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.541100 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.541206 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.588720 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4"] Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.645449 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.645511 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slx5t\" (UniqueName: \"kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.646036 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.737398 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.737639 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.747911 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.747956 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slx5t\" (UniqueName: \"kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.748068 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.758848 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.760561 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.778533 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slx5t\" (UniqueName: \"kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.816901 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:23 crc kubenswrapper[4888]: I1124 00:55:23.903687 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:24 crc kubenswrapper[4888]: I1124 00:55:24.546657 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:24 crc kubenswrapper[4888]: I1124 00:55:24.607572 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4"] Nov 24 00:55:24 crc kubenswrapper[4888]: W1124 00:55:24.610000 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a52b97b_92bd_42ab_8b5b_80eafbdce0fa.slice/crio-6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea WatchSource:0}: Error finding container 6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea: Status 404 returned error can't find the container with id 6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea Nov 24 00:55:25 crc kubenswrapper[4888]: I1124 00:55:25.482359 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" event={"ID":"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa","Type":"ContainerStarted","Data":"4012e3bc40a0281cd50745adc810301ba39a7869517d93044d2bffaa830d9bcc"} Nov 24 00:55:25 crc kubenswrapper[4888]: I1124 00:55:25.482926 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" event={"ID":"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa","Type":"ContainerStarted","Data":"6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea"} Nov 24 00:55:25 crc kubenswrapper[4888]: I1124 00:55:25.531525 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" podStartSLOduration=2.104489794 podStartE2EDuration="2.531498204s" podCreationTimestamp="2025-11-24 00:55:23 +0000 UTC" firstStartedPulling="2025-11-24 00:55:24.61318692 +0000 UTC m=+1827.195870974" lastFinishedPulling="2025-11-24 00:55:25.04019533 +0000 UTC m=+1827.622879384" observedRunningTime="2025-11-24 00:55:25.510807733 +0000 UTC m=+1828.093491797" watchObservedRunningTime="2025-11-24 00:55:25.531498204 +0000 UTC m=+1828.114182258" Nov 24 00:55:26 crc kubenswrapper[4888]: I1124 00:55:26.972054 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:27 crc kubenswrapper[4888]: I1124 00:55:27.514386 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v4k6w" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="registry-server" containerID="cri-o://0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558" gracePeriod=2 Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.008441 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.059425 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh6zp\" (UniqueName: \"kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp\") pod \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.059498 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities\") pod \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.059541 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content\") pod \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\" (UID: \"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4\") " Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.061040 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities" (OuterVolumeSpecName: "utilities") pod "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" (UID: "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.072047 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp" (OuterVolumeSpecName: "kube-api-access-lh6zp") pod "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" (UID: "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4"). InnerVolumeSpecName "kube-api-access-lh6zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.077796 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" (UID: "fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.160789 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh6zp\" (UniqueName: \"kubernetes.io/projected/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-kube-api-access-lh6zp\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.160845 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.160855 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:28 crc kubenswrapper[4888]: E1124 00:55:28.294742 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc2033cd_a3c5_46ca_9a1e_67b0cec4efc4.slice\": RecentStats: unable to find data in memory cache]" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.537246 4888 generic.go:334] "Generic (PLEG): container finished" podID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerID="0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558" exitCode=0 Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.537298 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerDied","Data":"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558"} Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.537346 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4k6w" event={"ID":"fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4","Type":"ContainerDied","Data":"e0be21c6a1689a0d784f8d12c97741418c7698949a48c45036a26917e6513277"} Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.537371 4888 scope.go:117] "RemoveContainer" containerID="0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.537369 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4k6w" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.573693 4888 scope.go:117] "RemoveContainer" containerID="22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.581952 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.592966 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4k6w"] Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.613192 4888 scope.go:117] "RemoveContainer" containerID="7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.673539 4888 scope.go:117] "RemoveContainer" containerID="0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558" Nov 24 00:55:28 crc kubenswrapper[4888]: E1124 00:55:28.674153 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558\": container with ID starting with 0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558 not found: ID does not exist" containerID="0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.674231 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558"} err="failed to get container status \"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558\": rpc error: code = NotFound desc = could not find container \"0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558\": container with ID starting with 0b2b7de8a753af130343262207c409537aad58e303dac6bb5737110daac1b558 not found: ID does not exist" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.674261 4888 scope.go:117] "RemoveContainer" containerID="22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf" Nov 24 00:55:28 crc kubenswrapper[4888]: E1124 00:55:28.674845 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf\": container with ID starting with 22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf not found: ID does not exist" containerID="22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.674967 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf"} err="failed to get container status \"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf\": rpc error: code = NotFound desc = could not find container \"22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf\": container with ID starting with 22c76231b55ebf95e43c102ae5a24dab93d4c187fd3ce30a4805986432cc1ddf not found: ID does not exist" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.675072 4888 scope.go:117] "RemoveContainer" containerID="7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79" Nov 24 00:55:28 crc kubenswrapper[4888]: E1124 00:55:28.675741 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79\": container with ID starting with 7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79 not found: ID does not exist" containerID="7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79" Nov 24 00:55:28 crc kubenswrapper[4888]: I1124 00:55:28.675797 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79"} err="failed to get container status \"7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79\": rpc error: code = NotFound desc = could not find container \"7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79\": container with ID starting with 7ebecd3e903db578b7e950b466b4a8854776f0a08a9ea917a57ac6d95c80eb79 not found: ID does not exist" Nov 24 00:55:29 crc kubenswrapper[4888]: I1124 00:55:29.065465 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-pnrn8"] Nov 24 00:55:29 crc kubenswrapper[4888]: I1124 00:55:29.075522 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-pnrn8"] Nov 24 00:55:30 crc kubenswrapper[4888]: I1124 00:55:30.269100 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="922fc13b-8e48-4729-bf76-2244367cdce4" path="/var/lib/kubelet/pods/922fc13b-8e48-4729-bf76-2244367cdce4/volumes" Nov 24 00:55:30 crc kubenswrapper[4888]: I1124 00:55:30.283111 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" path="/var/lib/kubelet/pods/fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4/volumes" Nov 24 00:55:30 crc kubenswrapper[4888]: I1124 00:55:30.566306 4888 generic.go:334] "Generic (PLEG): container finished" podID="4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" containerID="4012e3bc40a0281cd50745adc810301ba39a7869517d93044d2bffaa830d9bcc" exitCode=0 Nov 24 00:55:30 crc kubenswrapper[4888]: I1124 00:55:30.566358 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" event={"ID":"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa","Type":"ContainerDied","Data":"4012e3bc40a0281cd50745adc810301ba39a7869517d93044d2bffaa830d9bcc"} Nov 24 00:55:31 crc kubenswrapper[4888]: I1124 00:55:31.625366 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:31 crc kubenswrapper[4888]: I1124 00:55:31.704621 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.107030 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.260837 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key\") pod \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.260949 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slx5t\" (UniqueName: \"kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t\") pod \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.261123 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory\") pod \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\" (UID: \"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa\") " Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.269428 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t" (OuterVolumeSpecName: "kube-api-access-slx5t") pod "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" (UID: "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa"). InnerVolumeSpecName "kube-api-access-slx5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.300087 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" (UID: "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.320015 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory" (OuterVolumeSpecName: "inventory") pod "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" (UID: "4a52b97b-92bd-42ab-8b5b-80eafbdce0fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.365290 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.365394 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.365408 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slx5t\" (UniqueName: \"kubernetes.io/projected/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa-kube-api-access-slx5t\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.377580 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.598109 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" event={"ID":"4a52b97b-92bd-42ab-8b5b-80eafbdce0fa","Type":"ContainerDied","Data":"6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea"} Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.598528 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ee897fba504761a4922f3a7cdc72d18bf1b49581138363f15aefd19096df4ea" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.598196 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.683848 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml"] Nov 24 00:55:32 crc kubenswrapper[4888]: E1124 00:55:32.684257 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="extract-utilities" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684271 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="extract-utilities" Nov 24 00:55:32 crc kubenswrapper[4888]: E1124 00:55:32.684280 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="extract-content" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684286 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="extract-content" Nov 24 00:55:32 crc kubenswrapper[4888]: E1124 00:55:32.684310 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684317 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:32 crc kubenswrapper[4888]: E1124 00:55:32.684349 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="registry-server" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684355 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="registry-server" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684559 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.684583 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2033cd-a3c5-46ca-9a1e-67b0cec4efc4" containerName="registry-server" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.685262 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.688046 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.688223 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.688590 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.688963 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.711288 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml"] Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.787920 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.788219 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.788874 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kh4j\" (UniqueName: \"kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.890654 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.890774 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kh4j\" (UniqueName: \"kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.890796 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.894757 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.905389 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:32 crc kubenswrapper[4888]: I1124 00:55:32.909193 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kh4j\" (UniqueName: \"kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sl6ml\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:33 crc kubenswrapper[4888]: I1124 00:55:33.061562 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:55:33 crc kubenswrapper[4888]: I1124 00:55:33.611702 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jstsv" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="registry-server" containerID="cri-o://950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72" gracePeriod=2 Nov 24 00:55:33 crc kubenswrapper[4888]: I1124 00:55:33.706471 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml"] Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.296094 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.424099 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content\") pod \"24bf2072-516b-4fc9-bec2-2b8910f1077b\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.424200 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7mt4\" (UniqueName: \"kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4\") pod \"24bf2072-516b-4fc9-bec2-2b8910f1077b\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.424409 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities\") pod \"24bf2072-516b-4fc9-bec2-2b8910f1077b\" (UID: \"24bf2072-516b-4fc9-bec2-2b8910f1077b\") " Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.425295 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities" (OuterVolumeSpecName: "utilities") pod "24bf2072-516b-4fc9-bec2-2b8910f1077b" (UID: "24bf2072-516b-4fc9-bec2-2b8910f1077b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.440974 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4" (OuterVolumeSpecName: "kube-api-access-k7mt4") pod "24bf2072-516b-4fc9-bec2-2b8910f1077b" (UID: "24bf2072-516b-4fc9-bec2-2b8910f1077b"). InnerVolumeSpecName "kube-api-access-k7mt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.530084 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.530310 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7mt4\" (UniqueName: \"kubernetes.io/projected/24bf2072-516b-4fc9-bec2-2b8910f1077b-kube-api-access-k7mt4\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.550564 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24bf2072-516b-4fc9-bec2-2b8910f1077b" (UID: "24bf2072-516b-4fc9-bec2-2b8910f1077b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.628703 4888 generic.go:334] "Generic (PLEG): container finished" podID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerID="950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72" exitCode=0 Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.628739 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerDied","Data":"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72"} Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.629155 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jstsv" event={"ID":"24bf2072-516b-4fc9-bec2-2b8910f1077b","Type":"ContainerDied","Data":"f61eb0ad258c3ab6ca273446a218fcafdc9d1dfe5dc59d44ee7be3f09b028494"} Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.628838 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jstsv" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.629180 4888 scope.go:117] "RemoveContainer" containerID="950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.633214 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24bf2072-516b-4fc9-bec2-2b8910f1077b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.634626 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" event={"ID":"52c6eda4-874e-479b-b8cc-2f9212a197f8","Type":"ContainerStarted","Data":"bc673cba94d2aa570b541e34c3178d7e6bcb000739417b668b26ec567bb817e0"} Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.634660 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" event={"ID":"52c6eda4-874e-479b-b8cc-2f9212a197f8","Type":"ContainerStarted","Data":"acbafd347c2abae0603b36a6d24e11763768e7b4961d68ae454b6bc8b46a155a"} Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.661111 4888 scope.go:117] "RemoveContainer" containerID="8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.661250 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" podStartSLOduration=2.217354116 podStartE2EDuration="2.661230509s" podCreationTimestamp="2025-11-24 00:55:32 +0000 UTC" firstStartedPulling="2025-11-24 00:55:33.715823226 +0000 UTC m=+1836.298507270" lastFinishedPulling="2025-11-24 00:55:34.159699619 +0000 UTC m=+1836.742383663" observedRunningTime="2025-11-24 00:55:34.654400998 +0000 UTC m=+1837.237085082" watchObservedRunningTime="2025-11-24 00:55:34.661230509 +0000 UTC m=+1837.243914553" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.686551 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.699425 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jstsv"] Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.712034 4888 scope.go:117] "RemoveContainer" containerID="65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.745036 4888 scope.go:117] "RemoveContainer" containerID="950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72" Nov 24 00:55:34 crc kubenswrapper[4888]: E1124 00:55:34.745698 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72\": container with ID starting with 950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72 not found: ID does not exist" containerID="950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.745751 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72"} err="failed to get container status \"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72\": rpc error: code = NotFound desc = could not find container \"950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72\": container with ID starting with 950ef385110c94586fbe2b09cc00172d3e1c3ede7e95920ed04003ae86ec4a72 not found: ID does not exist" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.745779 4888 scope.go:117] "RemoveContainer" containerID="8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5" Nov 24 00:55:34 crc kubenswrapper[4888]: E1124 00:55:34.746325 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5\": container with ID starting with 8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5 not found: ID does not exist" containerID="8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.746398 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5"} err="failed to get container status \"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5\": rpc error: code = NotFound desc = could not find container \"8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5\": container with ID starting with 8bc6b128d8bbb4c0e1c6387fac8ef4804235401ac5527b9fda66b187a32fd3d5 not found: ID does not exist" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.746460 4888 scope.go:117] "RemoveContainer" containerID="65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa" Nov 24 00:55:34 crc kubenswrapper[4888]: E1124 00:55:34.746873 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa\": container with ID starting with 65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa not found: ID does not exist" containerID="65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa" Nov 24 00:55:34 crc kubenswrapper[4888]: I1124 00:55:34.746905 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa"} err="failed to get container status \"65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa\": rpc error: code = NotFound desc = could not find container \"65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa\": container with ID starting with 65a013ccebe40ab92e33d4f646ef4708980816ad8e6611a33c85782ab02628aa not found: ID does not exist" Nov 24 00:55:36 crc kubenswrapper[4888]: I1124 00:55:36.267263 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" path="/var/lib/kubelet/pods/24bf2072-516b-4fc9-bec2-2b8910f1077b/volumes" Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.051007 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-w7rnb"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.069106 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-pkgvs"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.082326 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nh84n"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.094801 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-w7rnb"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.103989 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-pkgvs"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.112320 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nh84n"] Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.269789 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4132e22-3a19-46cf-b764-057325ee448d" path="/var/lib/kubelet/pods/a4132e22-3a19-46cf-b764-057325ee448d/volumes" Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.275772 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a76821af-c9e6-47a8-a167-1531770cc394" path="/var/lib/kubelet/pods/a76821af-c9e6-47a8-a167-1531770cc394/volumes" Nov 24 00:55:40 crc kubenswrapper[4888]: I1124 00:55:40.277332 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87929ab-dd6b-408f-80ed-e1716b5384dc" path="/var/lib/kubelet/pods/e87929ab-dd6b-408f-80ed-e1716b5384dc/volumes" Nov 24 00:55:59 crc kubenswrapper[4888]: I1124 00:55:59.045306 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jzk9c"] Nov 24 00:55:59 crc kubenswrapper[4888]: I1124 00:55:59.056091 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jzk9c"] Nov 24 00:56:00 crc kubenswrapper[4888]: I1124 00:56:00.257723 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5a8488-3848-4285-96bc-6313cc426ad5" path="/var/lib/kubelet/pods/1d5a8488-3848-4285-96bc-6313cc426ad5/volumes" Nov 24 00:56:10 crc kubenswrapper[4888]: I1124 00:56:10.491059 4888 scope.go:117] "RemoveContainer" containerID="b7efcef354ddd27e8794abf82259baf7138afcabf60301155aca700caafff68c" Nov 24 00:56:10 crc kubenswrapper[4888]: I1124 00:56:10.542369 4888 scope.go:117] "RemoveContainer" containerID="2c8e8dbb0ade978dd2c31689522314955b9fb1618cb0844ccfb122e1e44ee354" Nov 24 00:56:10 crc kubenswrapper[4888]: I1124 00:56:10.606225 4888 scope.go:117] "RemoveContainer" containerID="b97431cb0245a05e90ed1beeae629ae67ee41da52d41e9dd0163f46e7b1ecaff" Nov 24 00:56:10 crc kubenswrapper[4888]: I1124 00:56:10.677428 4888 scope.go:117] "RemoveContainer" containerID="e32ba11b3e32fdbb8345ec3932956ccfb6eec9e7d8e4068e72c2a106d8bc0631" Nov 24 00:56:10 crc kubenswrapper[4888]: I1124 00:56:10.726627 4888 scope.go:117] "RemoveContainer" containerID="7ddb707bef87a19cc9be039fbbf77a67ab65a2232d2d4f03a098472e123cd70e" Nov 24 00:56:16 crc kubenswrapper[4888]: I1124 00:56:16.753896 4888 generic.go:334] "Generic (PLEG): container finished" podID="52c6eda4-874e-479b-b8cc-2f9212a197f8" containerID="bc673cba94d2aa570b541e34c3178d7e6bcb000739417b668b26ec567bb817e0" exitCode=0 Nov 24 00:56:16 crc kubenswrapper[4888]: I1124 00:56:16.754374 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" event={"ID":"52c6eda4-874e-479b-b8cc-2f9212a197f8","Type":"ContainerDied","Data":"bc673cba94d2aa570b541e34c3178d7e6bcb000739417b668b26ec567bb817e0"} Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.345238 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.436133 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory\") pod \"52c6eda4-874e-479b-b8cc-2f9212a197f8\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.436195 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kh4j\" (UniqueName: \"kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j\") pod \"52c6eda4-874e-479b-b8cc-2f9212a197f8\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.436319 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key\") pod \"52c6eda4-874e-479b-b8cc-2f9212a197f8\" (UID: \"52c6eda4-874e-479b-b8cc-2f9212a197f8\") " Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.450122 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j" (OuterVolumeSpecName: "kube-api-access-5kh4j") pod "52c6eda4-874e-479b-b8cc-2f9212a197f8" (UID: "52c6eda4-874e-479b-b8cc-2f9212a197f8"). InnerVolumeSpecName "kube-api-access-5kh4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.471008 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory" (OuterVolumeSpecName: "inventory") pod "52c6eda4-874e-479b-b8cc-2f9212a197f8" (UID: "52c6eda4-874e-479b-b8cc-2f9212a197f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.488677 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52c6eda4-874e-479b-b8cc-2f9212a197f8" (UID: "52c6eda4-874e-479b-b8cc-2f9212a197f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.539373 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.539402 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kh4j\" (UniqueName: \"kubernetes.io/projected/52c6eda4-874e-479b-b8cc-2f9212a197f8-kube-api-access-5kh4j\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.539412 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52c6eda4-874e-479b-b8cc-2f9212a197f8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.790038 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" event={"ID":"52c6eda4-874e-479b-b8cc-2f9212a197f8","Type":"ContainerDied","Data":"acbafd347c2abae0603b36a6d24e11763768e7b4961d68ae454b6bc8b46a155a"} Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.790109 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acbafd347c2abae0603b36a6d24e11763768e7b4961d68ae454b6bc8b46a155a" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.790123 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.925987 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp"] Nov 24 00:56:18 crc kubenswrapper[4888]: E1124 00:56:18.926598 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="registry-server" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.926626 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="registry-server" Nov 24 00:56:18 crc kubenswrapper[4888]: E1124 00:56:18.926665 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c6eda4-874e-479b-b8cc-2f9212a197f8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.926679 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c6eda4-874e-479b-b8cc-2f9212a197f8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:18 crc kubenswrapper[4888]: E1124 00:56:18.926716 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="extract-utilities" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.926729 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="extract-utilities" Nov 24 00:56:18 crc kubenswrapper[4888]: E1124 00:56:18.926778 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="extract-content" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.926791 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="extract-content" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.927155 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="24bf2072-516b-4fc9-bec2-2b8910f1077b" containerName="registry-server" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.927192 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c6eda4-874e-479b-b8cc-2f9212a197f8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.928682 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.931910 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.931922 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.932854 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.933143 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:56:18 crc kubenswrapper[4888]: I1124 00:56:18.947476 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp"] Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.051205 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.051285 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqwx\" (UniqueName: \"kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.051369 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.152709 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.152826 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqwx\" (UniqueName: \"kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.152935 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.157151 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.159196 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.175072 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqwx\" (UniqueName: \"kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.250404 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:19 crc kubenswrapper[4888]: I1124 00:56:19.879585 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp"] Nov 24 00:56:20 crc kubenswrapper[4888]: I1124 00:56:20.813393 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" event={"ID":"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9","Type":"ContainerStarted","Data":"70fcba0c6aa150fb3696f379be7fdab98b6cd1006b50926a9b2b1339ca6610c8"} Nov 24 00:56:20 crc kubenswrapper[4888]: I1124 00:56:20.813776 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" event={"ID":"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9","Type":"ContainerStarted","Data":"8c404b5c795c854265b1fb75e4d8fb6d27b79ffb142ce77816c1932955ff9f40"} Nov 24 00:56:20 crc kubenswrapper[4888]: I1124 00:56:20.848453 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" podStartSLOduration=2.406964214 podStartE2EDuration="2.848431149s" podCreationTimestamp="2025-11-24 00:56:18 +0000 UTC" firstStartedPulling="2025-11-24 00:56:19.89896316 +0000 UTC m=+1882.481647204" lastFinishedPulling="2025-11-24 00:56:20.340430045 +0000 UTC m=+1882.923114139" observedRunningTime="2025-11-24 00:56:20.828952662 +0000 UTC m=+1883.411636756" watchObservedRunningTime="2025-11-24 00:56:20.848431149 +0000 UTC m=+1883.431115203" Nov 24 00:56:25 crc kubenswrapper[4888]: I1124 00:56:25.876333 4888 generic.go:334] "Generic (PLEG): container finished" podID="4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" containerID="70fcba0c6aa150fb3696f379be7fdab98b6cd1006b50926a9b2b1339ca6610c8" exitCode=0 Nov 24 00:56:25 crc kubenswrapper[4888]: I1124 00:56:25.876438 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" event={"ID":"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9","Type":"ContainerDied","Data":"70fcba0c6aa150fb3696f379be7fdab98b6cd1006b50926a9b2b1339ca6610c8"} Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.371691 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.467415 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqqwx\" (UniqueName: \"kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx\") pod \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.467517 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory\") pod \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.467670 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key\") pod \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\" (UID: \"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9\") " Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.477598 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx" (OuterVolumeSpecName: "kube-api-access-fqqwx") pod "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" (UID: "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9"). InnerVolumeSpecName "kube-api-access-fqqwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.510870 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" (UID: "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.517293 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory" (OuterVolumeSpecName: "inventory") pod "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" (UID: "4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.570952 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.571014 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.571025 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqqwx\" (UniqueName: \"kubernetes.io/projected/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9-kube-api-access-fqqwx\") on node \"crc\" DevicePath \"\"" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.918208 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" event={"ID":"4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9","Type":"ContainerDied","Data":"8c404b5c795c854265b1fb75e4d8fb6d27b79ffb142ce77816c1932955ff9f40"} Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.918266 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c404b5c795c854265b1fb75e4d8fb6d27b79ffb142ce77816c1932955ff9f40" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.918397 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.978286 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m"] Nov 24 00:56:27 crc kubenswrapper[4888]: E1124 00:56:27.979177 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.979199 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.979422 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.980359 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.984029 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.984143 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.984238 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:56:27 crc kubenswrapper[4888]: I1124 00:56:27.984400 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:27.997671 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m"] Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.081236 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.081494 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzsdq\" (UniqueName: \"kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.081627 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.184040 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.184214 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzsdq\" (UniqueName: \"kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.184373 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.188773 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.193742 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.212494 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzsdq\" (UniqueName: \"kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-74w7m\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.314708 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.660763 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m"] Nov 24 00:56:28 crc kubenswrapper[4888]: I1124 00:56:28.927082 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" event={"ID":"67f5d5b8-4bdd-4601-978f-30a1dff053c0","Type":"ContainerStarted","Data":"97fc4b0a023fbb75f1184e63a1d16e6501f380687a82806f765232a9c7c21c27"} Nov 24 00:56:29 crc kubenswrapper[4888]: I1124 00:56:29.956399 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" event={"ID":"67f5d5b8-4bdd-4601-978f-30a1dff053c0","Type":"ContainerStarted","Data":"92af64c26868dc23429e73d4be9e712469f06ed7274b2a50f9b2c5394427ad0e"} Nov 24 00:56:29 crc kubenswrapper[4888]: I1124 00:56:29.998771 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" podStartSLOduration=2.532986804 podStartE2EDuration="2.998740591s" podCreationTimestamp="2025-11-24 00:56:27 +0000 UTC" firstStartedPulling="2025-11-24 00:56:28.665919519 +0000 UTC m=+1891.248603573" lastFinishedPulling="2025-11-24 00:56:29.131673306 +0000 UTC m=+1891.714357360" observedRunningTime="2025-11-24 00:56:29.990894731 +0000 UTC m=+1892.573578775" watchObservedRunningTime="2025-11-24 00:56:29.998740591 +0000 UTC m=+1892.581424675" Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.052299 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a6aa-account-create-25cm2"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.065755 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a6aa-account-create-25cm2"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.077164 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f706-account-create-p74tv"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.087191 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7kgw5"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.094956 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7kgw5"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.102322 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8291-account-create-mpq4l"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.109634 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f706-account-create-p74tv"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.117550 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-krvgq"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.124912 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-n7966"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.132107 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-n7966"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.139113 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-krvgq"] Nov 24 00:56:51 crc kubenswrapper[4888]: I1124 00:56:51.146415 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8291-account-create-mpq4l"] Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.270328 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fa347cb-768e-4e7b-ba47-42f178dec4ab" path="/var/lib/kubelet/pods/4fa347cb-768e-4e7b-ba47-42f178dec4ab/volumes" Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.272746 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b6f3b1-832a-4340-92dd-55b7e2797d49" path="/var/lib/kubelet/pods/59b6f3b1-832a-4340-92dd-55b7e2797d49/volumes" Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.276047 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74731cdc-25ca-44c5-a01b-ebe728556384" path="/var/lib/kubelet/pods/74731cdc-25ca-44c5-a01b-ebe728556384/volumes" Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.278777 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df146d80-370a-4eac-afb5-35e747ed7007" path="/var/lib/kubelet/pods/df146d80-370a-4eac-afb5-35e747ed7007/volumes" Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.281974 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18c5664-ccfc-4ae6-a9c2-5a543cca5d61" path="/var/lib/kubelet/pods/f18c5664-ccfc-4ae6-a9c2-5a543cca5d61/volumes" Nov 24 00:56:52 crc kubenswrapper[4888]: I1124 00:56:52.284507 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa3acc5c-2554-4d9c-b515-b01a8f0fbd21" path="/var/lib/kubelet/pods/fa3acc5c-2554-4d9c-b515-b01a8f0fbd21/volumes" Nov 24 00:57:10 crc kubenswrapper[4888]: I1124 00:57:10.893258 4888 scope.go:117] "RemoveContainer" containerID="024131d8514d8124a940492cb306d59b98c899da4cb6288834a4ff72f8343f26" Nov 24 00:57:10 crc kubenswrapper[4888]: I1124 00:57:10.938750 4888 scope.go:117] "RemoveContainer" containerID="8aa4efcffd04b0341bb00d54dc7bf77d11fd47c9503b5aaac3cf56eaabaabdf5" Nov 24 00:57:10 crc kubenswrapper[4888]: I1124 00:57:10.986442 4888 scope.go:117] "RemoveContainer" containerID="2acd0defe8157d1daac8851fd7575f316bebaf11336fc50a578fb04d7ab5f947" Nov 24 00:57:11 crc kubenswrapper[4888]: I1124 00:57:11.032973 4888 scope.go:117] "RemoveContainer" containerID="a963745166ee4f2cb69db0b8cb83c9697de979b3e2e9637074edce4745a29b42" Nov 24 00:57:11 crc kubenswrapper[4888]: I1124 00:57:11.076863 4888 scope.go:117] "RemoveContainer" containerID="5742acb8bd64d8d25648c70f6547a157a1e55fcb6c81f42f029b325309b57aad" Nov 24 00:57:11 crc kubenswrapper[4888]: I1124 00:57:11.148802 4888 scope.go:117] "RemoveContainer" containerID="79326d578c816a9c3b9bd2072e82d6f94e02c722d967770f2dc437b004d4e021" Nov 24 00:57:16 crc kubenswrapper[4888]: I1124 00:57:16.078244 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fz6k2"] Nov 24 00:57:16 crc kubenswrapper[4888]: I1124 00:57:16.098440 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fz6k2"] Nov 24 00:57:16 crc kubenswrapper[4888]: I1124 00:57:16.266741 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="141329ed-8501-47c4-bf4e-02bcde53a932" path="/var/lib/kubelet/pods/141329ed-8501-47c4-bf4e-02bcde53a932/volumes" Nov 24 00:57:23 crc kubenswrapper[4888]: I1124 00:57:23.591543 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:57:23 crc kubenswrapper[4888]: I1124 00:57:23.592088 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:57:25 crc kubenswrapper[4888]: I1124 00:57:25.670385 4888 generic.go:334] "Generic (PLEG): container finished" podID="67f5d5b8-4bdd-4601-978f-30a1dff053c0" containerID="92af64c26868dc23429e73d4be9e712469f06ed7274b2a50f9b2c5394427ad0e" exitCode=0 Nov 24 00:57:25 crc kubenswrapper[4888]: I1124 00:57:25.670887 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" event={"ID":"67f5d5b8-4bdd-4601-978f-30a1dff053c0","Type":"ContainerDied","Data":"92af64c26868dc23429e73d4be9e712469f06ed7274b2a50f9b2c5394427ad0e"} Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.043581 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-a098-account-create-7tggp"] Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.070203 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-fx726"] Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.081174 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-fx726"] Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.095069 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-a098-account-create-7tggp"] Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.256814 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.347953 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory\") pod \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.348109 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key\") pod \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.348208 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzsdq\" (UniqueName: \"kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq\") pod \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\" (UID: \"67f5d5b8-4bdd-4601-978f-30a1dff053c0\") " Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.356093 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq" (OuterVolumeSpecName: "kube-api-access-fzsdq") pod "67f5d5b8-4bdd-4601-978f-30a1dff053c0" (UID: "67f5d5b8-4bdd-4601-978f-30a1dff053c0"). InnerVolumeSpecName "kube-api-access-fzsdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.388338 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory" (OuterVolumeSpecName: "inventory") pod "67f5d5b8-4bdd-4601-978f-30a1dff053c0" (UID: "67f5d5b8-4bdd-4601-978f-30a1dff053c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.397959 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67f5d5b8-4bdd-4601-978f-30a1dff053c0" (UID: "67f5d5b8-4bdd-4601-978f-30a1dff053c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.450732 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.450769 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f5d5b8-4bdd-4601-978f-30a1dff053c0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.450782 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzsdq\" (UniqueName: \"kubernetes.io/projected/67f5d5b8-4bdd-4601-978f-30a1dff053c0-kube-api-access-fzsdq\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.700697 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" event={"ID":"67f5d5b8-4bdd-4601-978f-30a1dff053c0","Type":"ContainerDied","Data":"97fc4b0a023fbb75f1184e63a1d16e6501f380687a82806f765232a9c7c21c27"} Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.700736 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97fc4b0a023fbb75f1184e63a1d16e6501f380687a82806f765232a9c7c21c27" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.700806 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.800185 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qdqdf"] Nov 24 00:57:27 crc kubenswrapper[4888]: E1124 00:57:27.800774 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f5d5b8-4bdd-4601-978f-30a1dff053c0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.800806 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f5d5b8-4bdd-4601-978f-30a1dff053c0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.801222 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="67f5d5b8-4bdd-4601-978f-30a1dff053c0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.802184 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.805972 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.806203 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.809673 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.810099 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.830922 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qdqdf"] Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.859033 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.859148 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrjt6\" (UniqueName: \"kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.859228 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.960755 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrjt6\" (UniqueName: \"kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.960965 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.961091 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.964710 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.972481 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:27 crc kubenswrapper[4888]: I1124 00:57:27.977600 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrjt6\" (UniqueName: \"kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6\") pod \"ssh-known-hosts-edpm-deployment-qdqdf\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:28 crc kubenswrapper[4888]: I1124 00:57:28.146785 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:28 crc kubenswrapper[4888]: I1124 00:57:28.267358 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43af5b1c-1c23-41ea-b90a-f2025b4bb8ed" path="/var/lib/kubelet/pods/43af5b1c-1c23-41ea-b90a-f2025b4bb8ed/volumes" Nov 24 00:57:28 crc kubenswrapper[4888]: I1124 00:57:28.268189 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b23b2a93-fe3a-4fff-a78e-13a50826aefb" path="/var/lib/kubelet/pods/b23b2a93-fe3a-4fff-a78e-13a50826aefb/volumes" Nov 24 00:57:28 crc kubenswrapper[4888]: I1124 00:57:28.790961 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qdqdf"] Nov 24 00:57:29 crc kubenswrapper[4888]: I1124 00:57:29.725223 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" event={"ID":"089fe5dd-64b5-4982-aa29-4c8025a04581","Type":"ContainerStarted","Data":"89489d7a4ece5c2081b8ce17e9e3c7a1a376579875933344d8b27632e73194c7"} Nov 24 00:57:29 crc kubenswrapper[4888]: I1124 00:57:29.725557 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" event={"ID":"089fe5dd-64b5-4982-aa29-4c8025a04581","Type":"ContainerStarted","Data":"522f09b860c3d08cba0ba37777de181dc314d435e3ead1685de0f0c493c712c5"} Nov 24 00:57:29 crc kubenswrapper[4888]: I1124 00:57:29.741761 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" podStartSLOduration=2.278014543 podStartE2EDuration="2.741736361s" podCreationTimestamp="2025-11-24 00:57:27 +0000 UTC" firstStartedPulling="2025-11-24 00:57:28.793749507 +0000 UTC m=+1951.376433541" lastFinishedPulling="2025-11-24 00:57:29.257471315 +0000 UTC m=+1951.840155359" observedRunningTime="2025-11-24 00:57:29.739470967 +0000 UTC m=+1952.322155021" watchObservedRunningTime="2025-11-24 00:57:29.741736361 +0000 UTC m=+1952.324420405" Nov 24 00:57:37 crc kubenswrapper[4888]: I1124 00:57:37.817232 4888 generic.go:334] "Generic (PLEG): container finished" podID="089fe5dd-64b5-4982-aa29-4c8025a04581" containerID="89489d7a4ece5c2081b8ce17e9e3c7a1a376579875933344d8b27632e73194c7" exitCode=0 Nov 24 00:57:37 crc kubenswrapper[4888]: I1124 00:57:37.817341 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" event={"ID":"089fe5dd-64b5-4982-aa29-4c8025a04581","Type":"ContainerDied","Data":"89489d7a4ece5c2081b8ce17e9e3c7a1a376579875933344d8b27632e73194c7"} Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.377903 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.529722 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam\") pod \"089fe5dd-64b5-4982-aa29-4c8025a04581\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.529885 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0\") pod \"089fe5dd-64b5-4982-aa29-4c8025a04581\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.529974 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrjt6\" (UniqueName: \"kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6\") pod \"089fe5dd-64b5-4982-aa29-4c8025a04581\" (UID: \"089fe5dd-64b5-4982-aa29-4c8025a04581\") " Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.536572 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6" (OuterVolumeSpecName: "kube-api-access-qrjt6") pod "089fe5dd-64b5-4982-aa29-4c8025a04581" (UID: "089fe5dd-64b5-4982-aa29-4c8025a04581"). InnerVolumeSpecName "kube-api-access-qrjt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.560753 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "089fe5dd-64b5-4982-aa29-4c8025a04581" (UID: "089fe5dd-64b5-4982-aa29-4c8025a04581"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.561080 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "089fe5dd-64b5-4982-aa29-4c8025a04581" (UID: "089fe5dd-64b5-4982-aa29-4c8025a04581"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.633904 4888 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.633976 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrjt6\" (UniqueName: \"kubernetes.io/projected/089fe5dd-64b5-4982-aa29-4c8025a04581-kube-api-access-qrjt6\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.634007 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/089fe5dd-64b5-4982-aa29-4c8025a04581-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.844128 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" event={"ID":"089fe5dd-64b5-4982-aa29-4c8025a04581","Type":"ContainerDied","Data":"522f09b860c3d08cba0ba37777de181dc314d435e3ead1685de0f0c493c712c5"} Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.844189 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="522f09b860c3d08cba0ba37777de181dc314d435e3ead1685de0f0c493c712c5" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.844207 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qdqdf" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.925012 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq"] Nov 24 00:57:39 crc kubenswrapper[4888]: E1124 00:57:39.925794 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089fe5dd-64b5-4982-aa29-4c8025a04581" containerName="ssh-known-hosts-edpm-deployment" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.925848 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="089fe5dd-64b5-4982-aa29-4c8025a04581" containerName="ssh-known-hosts-edpm-deployment" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.926232 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="089fe5dd-64b5-4982-aa29-4c8025a04581" containerName="ssh-known-hosts-edpm-deployment" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.927806 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.930081 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.930643 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.931041 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.931062 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq"] Nov 24 00:57:39 crc kubenswrapper[4888]: I1124 00:57:39.931641 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.027647 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vd92r"] Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.038516 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vd92r"] Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.041297 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.041380 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvbr\" (UniqueName: \"kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.041461 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.143311 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.143390 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvbr\" (UniqueName: \"kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.143458 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.147791 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.148314 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.169598 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvbr\" (UniqueName: \"kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zjdzq\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.258501 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a3e9761-6aaf-43dd-ac54-d67a186b47b6" path="/var/lib/kubelet/pods/1a3e9761-6aaf-43dd-ac54-d67a186b47b6/volumes" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.259497 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.815340 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq"] Nov 24 00:57:40 crc kubenswrapper[4888]: I1124 00:57:40.854029 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" event={"ID":"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3","Type":"ContainerStarted","Data":"13f600ce6b894e68e9261f89e3a32a8ff5005be71080d4aa7621c51cf891224e"} Nov 24 00:57:41 crc kubenswrapper[4888]: I1124 00:57:41.868378 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" event={"ID":"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3","Type":"ContainerStarted","Data":"0c936c71c2b08c6db064838ebc13a00e9a15210df527ffcc121fa5dc71395dc1"} Nov 24 00:57:43 crc kubenswrapper[4888]: I1124 00:57:43.025794 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" podStartSLOduration=3.565063344 podStartE2EDuration="4.025778458s" podCreationTimestamp="2025-11-24 00:57:39 +0000 UTC" firstStartedPulling="2025-11-24 00:57:40.826413606 +0000 UTC m=+1963.409097670" lastFinishedPulling="2025-11-24 00:57:41.28712874 +0000 UTC m=+1963.869812784" observedRunningTime="2025-11-24 00:57:41.891204152 +0000 UTC m=+1964.473888226" watchObservedRunningTime="2025-11-24 00:57:43.025778458 +0000 UTC m=+1965.608462502" Nov 24 00:57:43 crc kubenswrapper[4888]: I1124 00:57:43.033784 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgpfg"] Nov 24 00:57:43 crc kubenswrapper[4888]: I1124 00:57:43.046250 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgpfg"] Nov 24 00:57:44 crc kubenswrapper[4888]: I1124 00:57:44.261039 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02a71bf4-35a2-4442-b9e5-828e553ad5e1" path="/var/lib/kubelet/pods/02a71bf4-35a2-4442-b9e5-828e553ad5e1/volumes" Nov 24 00:57:50 crc kubenswrapper[4888]: I1124 00:57:50.975477 4888 generic.go:334] "Generic (PLEG): container finished" podID="3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" containerID="0c936c71c2b08c6db064838ebc13a00e9a15210df527ffcc121fa5dc71395dc1" exitCode=0 Nov 24 00:57:50 crc kubenswrapper[4888]: I1124 00:57:50.975607 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" event={"ID":"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3","Type":"ContainerDied","Data":"0c936c71c2b08c6db064838ebc13a00e9a15210df527ffcc121fa5dc71395dc1"} Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.578535 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.726385 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory\") pod \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.726497 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key\") pod \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.726525 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzvbr\" (UniqueName: \"kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr\") pod \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\" (UID: \"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3\") " Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.731598 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr" (OuterVolumeSpecName: "kube-api-access-xzvbr") pod "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" (UID: "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3"). InnerVolumeSpecName "kube-api-access-xzvbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.752887 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory" (OuterVolumeSpecName: "inventory") pod "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" (UID: "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.785049 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" (UID: "3f89ceb2-718c-4ae5-bd9c-7b518d550cd3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.828658 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.828843 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:52 crc kubenswrapper[4888]: I1124 00:57:52.828903 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzvbr\" (UniqueName: \"kubernetes.io/projected/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3-kube-api-access-xzvbr\") on node \"crc\" DevicePath \"\"" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.011278 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" event={"ID":"3f89ceb2-718c-4ae5-bd9c-7b518d550cd3","Type":"ContainerDied","Data":"13f600ce6b894e68e9261f89e3a32a8ff5005be71080d4aa7621c51cf891224e"} Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.011323 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13f600ce6b894e68e9261f89e3a32a8ff5005be71080d4aa7621c51cf891224e" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.011378 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.123895 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj"] Nov 24 00:57:53 crc kubenswrapper[4888]: E1124 00:57:53.124275 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.124292 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.124517 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.125301 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.132477 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.132523 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.132726 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.132959 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.141933 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj"] Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.236181 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6tqh\" (UniqueName: \"kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.236247 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.236686 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.338450 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.338613 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.338716 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6tqh\" (UniqueName: \"kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.344513 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.344526 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.357614 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6tqh\" (UniqueName: \"kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.490127 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.592515 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:57:53 crc kubenswrapper[4888]: I1124 00:57:53.592613 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:57:54 crc kubenswrapper[4888]: I1124 00:57:54.077055 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj"] Nov 24 00:57:55 crc kubenswrapper[4888]: I1124 00:57:55.048789 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" event={"ID":"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b","Type":"ContainerStarted","Data":"de61341608f4f48f79daf8dc32d4f5cfd4e0e323f804295b47ed98a67e17f959"} Nov 24 00:57:55 crc kubenswrapper[4888]: I1124 00:57:55.049250 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" event={"ID":"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b","Type":"ContainerStarted","Data":"21cc38bdffab35f48a1eca4d7b1d75c1a618e0c87f860e2b0d7233057dd8c1b2"} Nov 24 00:57:55 crc kubenswrapper[4888]: I1124 00:57:55.096004 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" podStartSLOduration=1.645023375 podStartE2EDuration="2.095975013s" podCreationTimestamp="2025-11-24 00:57:53 +0000 UTC" firstStartedPulling="2025-11-24 00:57:54.091194225 +0000 UTC m=+1976.673878269" lastFinishedPulling="2025-11-24 00:57:54.542145863 +0000 UTC m=+1977.124829907" observedRunningTime="2025-11-24 00:57:55.076754213 +0000 UTC m=+1977.659438327" watchObservedRunningTime="2025-11-24 00:57:55.095975013 +0000 UTC m=+1977.678659097" Nov 24 00:58:05 crc kubenswrapper[4888]: I1124 00:58:05.174005 4888 generic.go:334] "Generic (PLEG): container finished" podID="1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" containerID="de61341608f4f48f79daf8dc32d4f5cfd4e0e323f804295b47ed98a67e17f959" exitCode=0 Nov 24 00:58:05 crc kubenswrapper[4888]: I1124 00:58:05.174138 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" event={"ID":"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b","Type":"ContainerDied","Data":"de61341608f4f48f79daf8dc32d4f5cfd4e0e323f804295b47ed98a67e17f959"} Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.710506 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.800798 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key\") pod \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.801394 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory\") pod \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.801709 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6tqh\" (UniqueName: \"kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh\") pod \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\" (UID: \"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b\") " Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.810349 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh" (OuterVolumeSpecName: "kube-api-access-z6tqh") pod "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" (UID: "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b"). InnerVolumeSpecName "kube-api-access-z6tqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.840881 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory" (OuterVolumeSpecName: "inventory") pod "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" (UID: "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.849100 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" (UID: "1d8f2d24-17cd-4aed-b5f3-b71ef305b58b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.915461 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6tqh\" (UniqueName: \"kubernetes.io/projected/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-kube-api-access-z6tqh\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.915898 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:06 crc kubenswrapper[4888]: I1124 00:58:06.915948 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.199310 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" event={"ID":"1d8f2d24-17cd-4aed-b5f3-b71ef305b58b","Type":"ContainerDied","Data":"21cc38bdffab35f48a1eca4d7b1d75c1a618e0c87f860e2b0d7233057dd8c1b2"} Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.199597 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21cc38bdffab35f48a1eca4d7b1d75c1a618e0c87f860e2b0d7233057dd8c1b2" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.199395 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.313583 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr"] Nov 24 00:58:07 crc kubenswrapper[4888]: E1124 00:58:07.314491 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.314635 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.315108 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.316350 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.322173 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.322508 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.322902 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.323272 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.323534 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.324036 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.324498 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.327909 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.354483 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr"] Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429191 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429679 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429743 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429874 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429954 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.429987 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430025 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430062 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430091 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430184 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430245 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430276 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tfh8\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.430419 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533036 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533089 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533129 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533167 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533190 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533221 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533250 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533274 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533354 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533399 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533432 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tfh8\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533471 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.533917 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.541964 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.542020 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.542053 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.543299 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.544206 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.544782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.545393 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.546445 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.546612 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.547943 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.548661 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.549724 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.557910 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tfh8\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:07 crc kubenswrapper[4888]: I1124 00:58:07.641798 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:08 crc kubenswrapper[4888]: I1124 00:58:08.266554 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr"] Nov 24 00:58:09 crc kubenswrapper[4888]: I1124 00:58:09.232389 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" event={"ID":"a90f3597-c206-488b-a833-49c9dd906ac9","Type":"ContainerStarted","Data":"1bb0be91508bf755e4688fb9709896762542e4c6ef528b6e98303296dc504159"} Nov 24 00:58:09 crc kubenswrapper[4888]: I1124 00:58:09.232999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" event={"ID":"a90f3597-c206-488b-a833-49c9dd906ac9","Type":"ContainerStarted","Data":"e7453e3c5570e6db3959940a4bbfad59e5f1d43d42c1fb0021a8ae7cd262a9ae"} Nov 24 00:58:09 crc kubenswrapper[4888]: I1124 00:58:09.255141 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" podStartSLOduration=1.820306013 podStartE2EDuration="2.255110089s" podCreationTimestamp="2025-11-24 00:58:07 +0000 UTC" firstStartedPulling="2025-11-24 00:58:08.273287845 +0000 UTC m=+1990.855971899" lastFinishedPulling="2025-11-24 00:58:08.708091891 +0000 UTC m=+1991.290775975" observedRunningTime="2025-11-24 00:58:09.251012694 +0000 UTC m=+1991.833696778" watchObservedRunningTime="2025-11-24 00:58:09.255110089 +0000 UTC m=+1991.837794163" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.307127 4888 scope.go:117] "RemoveContainer" containerID="9b2bae8aaffde05723f00f49fae9acf262fd364914738f71f574e40a08039c11" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.361969 4888 scope.go:117] "RemoveContainer" containerID="e896c578b132a12412e9a49f777d8686b34a6d316ffe3e72d39f23387f5806c7" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.429453 4888 scope.go:117] "RemoveContainer" containerID="0a19c69bb1abb58b20426b7b81a3e6f3de3a6aa5a8d50a4bbbe40fdabf8bfafd" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.479520 4888 scope.go:117] "RemoveContainer" containerID="f6a6deafce7428a5d422cebe9f3e9791ab247bf8374d123870a82a2198d25d56" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.543734 4888 scope.go:117] "RemoveContainer" containerID="6f7dd1280ac78f8578d663849476be4e1ae40e0cabb40af6ab076efaa519f3d8" Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.983269 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:11 crc kubenswrapper[4888]: I1124 00:58:11.988440 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.009134 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.054012 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mdhq\" (UniqueName: \"kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.054405 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.054573 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.157193 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mdhq\" (UniqueName: \"kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.157253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.157313 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.157900 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.158125 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.177965 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mdhq\" (UniqueName: \"kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq\") pod \"certified-operators-xg7ss\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.353342 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:12 crc kubenswrapper[4888]: I1124 00:58:12.876214 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:12 crc kubenswrapper[4888]: W1124 00:58:12.879756 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e2b2a5_95b3_4846_8563_093725d21d7e.slice/crio-9946fa463862932c44fe2d499a56ffb7329bbe48dc8a61518580a5a9488770f6 WatchSource:0}: Error finding container 9946fa463862932c44fe2d499a56ffb7329bbe48dc8a61518580a5a9488770f6: Status 404 returned error can't find the container with id 9946fa463862932c44fe2d499a56ffb7329bbe48dc8a61518580a5a9488770f6 Nov 24 00:58:13 crc kubenswrapper[4888]: E1124 00:58:13.258471 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e2b2a5_95b3_4846_8563_093725d21d7e.slice/crio-conmon-04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223.scope\": RecentStats: unable to find data in memory cache]" Nov 24 00:58:13 crc kubenswrapper[4888]: I1124 00:58:13.292168 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerID="04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223" exitCode=0 Nov 24 00:58:13 crc kubenswrapper[4888]: I1124 00:58:13.292211 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerDied","Data":"04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223"} Nov 24 00:58:13 crc kubenswrapper[4888]: I1124 00:58:13.292248 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerStarted","Data":"9946fa463862932c44fe2d499a56ffb7329bbe48dc8a61518580a5a9488770f6"} Nov 24 00:58:14 crc kubenswrapper[4888]: I1124 00:58:14.304908 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerStarted","Data":"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25"} Nov 24 00:58:16 crc kubenswrapper[4888]: I1124 00:58:16.331079 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerID="00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25" exitCode=0 Nov 24 00:58:16 crc kubenswrapper[4888]: I1124 00:58:16.331161 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerDied","Data":"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25"} Nov 24 00:58:17 crc kubenswrapper[4888]: I1124 00:58:17.351124 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerStarted","Data":"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6"} Nov 24 00:58:17 crc kubenswrapper[4888]: I1124 00:58:17.399725 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xg7ss" podStartSLOduration=2.948120011 podStartE2EDuration="6.399698923s" podCreationTimestamp="2025-11-24 00:58:11 +0000 UTC" firstStartedPulling="2025-11-24 00:58:13.294754974 +0000 UTC m=+1995.877439018" lastFinishedPulling="2025-11-24 00:58:16.746333886 +0000 UTC m=+1999.329017930" observedRunningTime="2025-11-24 00:58:17.379183167 +0000 UTC m=+1999.961867251" watchObservedRunningTime="2025-11-24 00:58:17.399698923 +0000 UTC m=+1999.982382977" Nov 24 00:58:22 crc kubenswrapper[4888]: I1124 00:58:22.353474 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:22 crc kubenswrapper[4888]: I1124 00:58:22.354093 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:22 crc kubenswrapper[4888]: I1124 00:58:22.418951 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:22 crc kubenswrapper[4888]: I1124 00:58:22.491108 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:22 crc kubenswrapper[4888]: I1124 00:58:22.661405 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:23 crc kubenswrapper[4888]: I1124 00:58:23.591720 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 00:58:23 crc kubenswrapper[4888]: I1124 00:58:23.592012 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 00:58:23 crc kubenswrapper[4888]: I1124 00:58:23.592056 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 00:58:23 crc kubenswrapper[4888]: I1124 00:58:23.592746 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 00:58:23 crc kubenswrapper[4888]: I1124 00:58:23.592789 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63" gracePeriod=600 Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.445926 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63" exitCode=0 Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.446063 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63"} Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.446785 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1"} Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.446851 4888 scope.go:117] "RemoveContainer" containerID="d324a5f2423d41da5182fe3a8adda82e7533e925dfdac3ad21b4f15467cb9153" Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.447121 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xg7ss" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="registry-server" containerID="cri-o://99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6" gracePeriod=2 Nov 24 00:58:24 crc kubenswrapper[4888]: I1124 00:58:24.974279 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.054415 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content\") pod \"e7e2b2a5-95b3-4846-8563-093725d21d7e\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.054822 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities\") pod \"e7e2b2a5-95b3-4846-8563-093725d21d7e\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.055105 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mdhq\" (UniqueName: \"kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq\") pod \"e7e2b2a5-95b3-4846-8563-093725d21d7e\" (UID: \"e7e2b2a5-95b3-4846-8563-093725d21d7e\") " Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.056214 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities" (OuterVolumeSpecName: "utilities") pod "e7e2b2a5-95b3-4846-8563-093725d21d7e" (UID: "e7e2b2a5-95b3-4846-8563-093725d21d7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.065100 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq" (OuterVolumeSpecName: "kube-api-access-7mdhq") pod "e7e2b2a5-95b3-4846-8563-093725d21d7e" (UID: "e7e2b2a5-95b3-4846-8563-093725d21d7e"). InnerVolumeSpecName "kube-api-access-7mdhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.135170 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7e2b2a5-95b3-4846-8563-093725d21d7e" (UID: "e7e2b2a5-95b3-4846-8563-093725d21d7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.158789 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.158872 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e2b2a5-95b3-4846-8563-093725d21d7e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.158894 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mdhq\" (UniqueName: \"kubernetes.io/projected/e7e2b2a5-95b3-4846-8563-093725d21d7e-kube-api-access-7mdhq\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.459291 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerID="99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6" exitCode=0 Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.459387 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg7ss" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.459402 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerDied","Data":"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6"} Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.459481 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg7ss" event={"ID":"e7e2b2a5-95b3-4846-8563-093725d21d7e","Type":"ContainerDied","Data":"9946fa463862932c44fe2d499a56ffb7329bbe48dc8a61518580a5a9488770f6"} Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.459519 4888 scope.go:117] "RemoveContainer" containerID="99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.500434 4888 scope.go:117] "RemoveContainer" containerID="00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.501306 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.510861 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xg7ss"] Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.527152 4888 scope.go:117] "RemoveContainer" containerID="04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.587347 4888 scope.go:117] "RemoveContainer" containerID="99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6" Nov 24 00:58:25 crc kubenswrapper[4888]: E1124 00:58:25.587782 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6\": container with ID starting with 99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6 not found: ID does not exist" containerID="99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.587943 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6"} err="failed to get container status \"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6\": rpc error: code = NotFound desc = could not find container \"99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6\": container with ID starting with 99b1c03f45c6fd3eb7d560ad631ca5714db35ad6221d2ed47296a7c3d29591c6 not found: ID does not exist" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.587991 4888 scope.go:117] "RemoveContainer" containerID="00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25" Nov 24 00:58:25 crc kubenswrapper[4888]: E1124 00:58:25.588334 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25\": container with ID starting with 00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25 not found: ID does not exist" containerID="00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.588372 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25"} err="failed to get container status \"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25\": rpc error: code = NotFound desc = could not find container \"00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25\": container with ID starting with 00c47c73a05156771bde18a829b1124af4b3e3fd795b5f762319db4dab893c25 not found: ID does not exist" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.588397 4888 scope.go:117] "RemoveContainer" containerID="04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223" Nov 24 00:58:25 crc kubenswrapper[4888]: E1124 00:58:25.588657 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223\": container with ID starting with 04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223 not found: ID does not exist" containerID="04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223" Nov 24 00:58:25 crc kubenswrapper[4888]: I1124 00:58:25.588688 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223"} err="failed to get container status \"04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223\": rpc error: code = NotFound desc = could not find container \"04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223\": container with ID starting with 04437e1780945f89be95a394830f0167994febca63624c95d240dbabd5204223 not found: ID does not exist" Nov 24 00:58:26 crc kubenswrapper[4888]: I1124 00:58:26.260499 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" path="/var/lib/kubelet/pods/e7e2b2a5-95b3-4846-8563-093725d21d7e/volumes" Nov 24 00:58:29 crc kubenswrapper[4888]: I1124 00:58:29.054344 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmm2t"] Nov 24 00:58:29 crc kubenswrapper[4888]: I1124 00:58:29.064708 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmm2t"] Nov 24 00:58:30 crc kubenswrapper[4888]: I1124 00:58:30.268476 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab6f695-53c2-43fe-bef0-fea107068d30" path="/var/lib/kubelet/pods/cab6f695-53c2-43fe-bef0-fea107068d30/volumes" Nov 24 00:58:50 crc kubenswrapper[4888]: I1124 00:58:50.793669 4888 generic.go:334] "Generic (PLEG): container finished" podID="a90f3597-c206-488b-a833-49c9dd906ac9" containerID="1bb0be91508bf755e4688fb9709896762542e4c6ef528b6e98303296dc504159" exitCode=0 Nov 24 00:58:50 crc kubenswrapper[4888]: I1124 00:58:50.793954 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" event={"ID":"a90f3597-c206-488b-a833-49c9dd906ac9","Type":"ContainerDied","Data":"1bb0be91508bf755e4688fb9709896762542e4c6ef528b6e98303296dc504159"} Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.385176 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485504 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485749 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485786 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485857 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485874 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tfh8\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485904 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485953 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485973 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.485991 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.486022 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.486042 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.486072 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.486109 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a90f3597-c206-488b-a833-49c9dd906ac9\" (UID: \"a90f3597-c206-488b-a833-49c9dd906ac9\") " Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.492669 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495591 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495634 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.494968 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495259 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495668 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495489 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495571 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.495585 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.497070 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.504089 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8" (OuterVolumeSpecName: "kube-api-access-4tfh8") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "kube-api-access-4tfh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.519002 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory" (OuterVolumeSpecName: "inventory") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.526112 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a90f3597-c206-488b-a833-49c9dd906ac9" (UID: "a90f3597-c206-488b-a833-49c9dd906ac9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588083 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588123 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588138 4888 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588150 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588162 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tfh8\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-kube-api-access-4tfh8\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588173 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588182 4888 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588192 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588206 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588216 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588225 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588233 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90f3597-c206-488b-a833-49c9dd906ac9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.588241 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a90f3597-c206-488b-a833-49c9dd906ac9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.859888 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" event={"ID":"a90f3597-c206-488b-a833-49c9dd906ac9","Type":"ContainerDied","Data":"e7453e3c5570e6db3959940a4bbfad59e5f1d43d42c1fb0021a8ae7cd262a9ae"} Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.859959 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7453e3c5570e6db3959940a4bbfad59e5f1d43d42c1fb0021a8ae7cd262a9ae" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.860098 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.941640 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn"] Nov 24 00:58:52 crc kubenswrapper[4888]: E1124 00:58:52.942112 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="extract-utilities" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942129 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="extract-utilities" Nov 24 00:58:52 crc kubenswrapper[4888]: E1124 00:58:52.942143 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="extract-content" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942150 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="extract-content" Nov 24 00:58:52 crc kubenswrapper[4888]: E1124 00:58:52.942170 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90f3597-c206-488b-a833-49c9dd906ac9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942178 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90f3597-c206-488b-a833-49c9dd906ac9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:52 crc kubenswrapper[4888]: E1124 00:58:52.942210 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="registry-server" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942217 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="registry-server" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942407 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e2b2a5-95b3-4846-8563-093725d21d7e" containerName="registry-server" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.942437 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90f3597-c206-488b-a833-49c9dd906ac9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.943263 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.946790 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.946818 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.946921 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.947181 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.947236 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.953000 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn"] Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.996805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.996896 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6svxz\" (UniqueName: \"kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.996935 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.997002 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:52 crc kubenswrapper[4888]: I1124 00:58:52.997051 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.098108 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.098244 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.098292 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6svxz\" (UniqueName: \"kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.098326 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.098350 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.100002 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.103443 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.111840 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.117385 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.123875 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6svxz\" (UniqueName: \"kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tfbqn\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.273025 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 00:58:53 crc kubenswrapper[4888]: I1124 00:58:53.893513 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn"] Nov 24 00:58:54 crc kubenswrapper[4888]: I1124 00:58:54.887874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" event={"ID":"ca5b9705-eedc-46e0-9f41-4b070305f441","Type":"ContainerStarted","Data":"0ece25aa8b97abdb3d29e524bf4bd33c7d0ca106dfc9f77fcdc10cbcfed948aa"} Nov 24 00:58:54 crc kubenswrapper[4888]: I1124 00:58:54.888369 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" event={"ID":"ca5b9705-eedc-46e0-9f41-4b070305f441","Type":"ContainerStarted","Data":"7a2f12c3ab7a890ebe52003a418da8a4b3250e5abb9f77a759285dd7d992087f"} Nov 24 00:58:54 crc kubenswrapper[4888]: I1124 00:58:54.915999 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" podStartSLOduration=2.3958448199999998 podStartE2EDuration="2.915966185s" podCreationTimestamp="2025-11-24 00:58:52 +0000 UTC" firstStartedPulling="2025-11-24 00:58:53.906363423 +0000 UTC m=+2036.489047467" lastFinishedPulling="2025-11-24 00:58:54.426484748 +0000 UTC m=+2037.009168832" observedRunningTime="2025-11-24 00:58:54.905669126 +0000 UTC m=+2037.488353220" watchObservedRunningTime="2025-11-24 00:58:54.915966185 +0000 UTC m=+2037.498650299" Nov 24 00:59:11 crc kubenswrapper[4888]: I1124 00:59:11.756573 4888 scope.go:117] "RemoveContainer" containerID="f03f887341c9e9500f0819e7e02b31aa263a5c14ba443321635f6bc69429f1d3" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.161707 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.164424 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.178214 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.264986 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r2kg\" (UniqueName: \"kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.265060 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.265127 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.368207 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r2kg\" (UniqueName: \"kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.368305 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.368481 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.369290 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.369336 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.388338 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r2kg\" (UniqueName: \"kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg\") pod \"redhat-operators-hhpsr\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:12 crc kubenswrapper[4888]: I1124 00:59:12.525530 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:13 crc kubenswrapper[4888]: I1124 00:59:13.016181 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:13 crc kubenswrapper[4888]: I1124 00:59:13.112856 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerStarted","Data":"79f880321df7e6a1f4224b5c950f61906b8b0d3b1361ca8e233bc22ac9b96205"} Nov 24 00:59:14 crc kubenswrapper[4888]: I1124 00:59:14.122701 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b06f61d-9a97-441b-808d-439551b418c5" containerID="f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea" exitCode=0 Nov 24 00:59:14 crc kubenswrapper[4888]: I1124 00:59:14.122858 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerDied","Data":"f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea"} Nov 24 00:59:15 crc kubenswrapper[4888]: I1124 00:59:15.139213 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerStarted","Data":"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452"} Nov 24 00:59:19 crc kubenswrapper[4888]: I1124 00:59:19.199065 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b06f61d-9a97-441b-808d-439551b418c5" containerID="d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452" exitCode=0 Nov 24 00:59:19 crc kubenswrapper[4888]: I1124 00:59:19.199127 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerDied","Data":"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452"} Nov 24 00:59:20 crc kubenswrapper[4888]: I1124 00:59:20.211874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerStarted","Data":"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774"} Nov 24 00:59:20 crc kubenswrapper[4888]: I1124 00:59:20.239837 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hhpsr" podStartSLOduration=2.763852183 podStartE2EDuration="8.239824137s" podCreationTimestamp="2025-11-24 00:59:12 +0000 UTC" firstStartedPulling="2025-11-24 00:59:14.124819429 +0000 UTC m=+2056.707503483" lastFinishedPulling="2025-11-24 00:59:19.600791383 +0000 UTC m=+2062.183475437" observedRunningTime="2025-11-24 00:59:20.233422078 +0000 UTC m=+2062.816106122" watchObservedRunningTime="2025-11-24 00:59:20.239824137 +0000 UTC m=+2062.822508181" Nov 24 00:59:22 crc kubenswrapper[4888]: I1124 00:59:22.526498 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:22 crc kubenswrapper[4888]: I1124 00:59:22.527792 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:23 crc kubenswrapper[4888]: I1124 00:59:23.578699 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hhpsr" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="registry-server" probeResult="failure" output=< Nov 24 00:59:23 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 00:59:23 crc kubenswrapper[4888]: > Nov 24 00:59:32 crc kubenswrapper[4888]: I1124 00:59:32.595041 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:32 crc kubenswrapper[4888]: I1124 00:59:32.669909 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:32 crc kubenswrapper[4888]: I1124 00:59:32.834699 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:34 crc kubenswrapper[4888]: I1124 00:59:34.363251 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hhpsr" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="registry-server" containerID="cri-o://b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774" gracePeriod=2 Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.021984 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.193024 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content\") pod \"5b06f61d-9a97-441b-808d-439551b418c5\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.193483 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities\") pod \"5b06f61d-9a97-441b-808d-439551b418c5\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.193725 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r2kg\" (UniqueName: \"kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg\") pod \"5b06f61d-9a97-441b-808d-439551b418c5\" (UID: \"5b06f61d-9a97-441b-808d-439551b418c5\") " Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.194365 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities" (OuterVolumeSpecName: "utilities") pod "5b06f61d-9a97-441b-808d-439551b418c5" (UID: "5b06f61d-9a97-441b-808d-439551b418c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.194579 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.200127 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg" (OuterVolumeSpecName: "kube-api-access-2r2kg") pod "5b06f61d-9a97-441b-808d-439551b418c5" (UID: "5b06f61d-9a97-441b-808d-439551b418c5"). InnerVolumeSpecName "kube-api-access-2r2kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.293077 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b06f61d-9a97-441b-808d-439551b418c5" (UID: "5b06f61d-9a97-441b-808d-439551b418c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.297507 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b06f61d-9a97-441b-808d-439551b418c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.297569 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r2kg\" (UniqueName: \"kubernetes.io/projected/5b06f61d-9a97-441b-808d-439551b418c5-kube-api-access-2r2kg\") on node \"crc\" DevicePath \"\"" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.378513 4888 generic.go:334] "Generic (PLEG): container finished" podID="5b06f61d-9a97-441b-808d-439551b418c5" containerID="b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774" exitCode=0 Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.378578 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerDied","Data":"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774"} Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.378620 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hhpsr" event={"ID":"5b06f61d-9a97-441b-808d-439551b418c5","Type":"ContainerDied","Data":"79f880321df7e6a1f4224b5c950f61906b8b0d3b1361ca8e233bc22ac9b96205"} Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.378653 4888 scope.go:117] "RemoveContainer" containerID="b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.378924 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hhpsr" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.441130 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.447110 4888 scope.go:117] "RemoveContainer" containerID="d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.448228 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hhpsr"] Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.484328 4888 scope.go:117] "RemoveContainer" containerID="f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.537062 4888 scope.go:117] "RemoveContainer" containerID="b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774" Nov 24 00:59:35 crc kubenswrapper[4888]: E1124 00:59:35.537430 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774\": container with ID starting with b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774 not found: ID does not exist" containerID="b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.537461 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774"} err="failed to get container status \"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774\": rpc error: code = NotFound desc = could not find container \"b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774\": container with ID starting with b8a592edc7a90522935b593c1181114c1d4047271099653c3e60a13fc6b75774 not found: ID does not exist" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.537480 4888 scope.go:117] "RemoveContainer" containerID="d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452" Nov 24 00:59:35 crc kubenswrapper[4888]: E1124 00:59:35.538412 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452\": container with ID starting with d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452 not found: ID does not exist" containerID="d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.538440 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452"} err="failed to get container status \"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452\": rpc error: code = NotFound desc = could not find container \"d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452\": container with ID starting with d7cf4a224fddf991b7fe24e08f856b4e572783aee1aab66041897f6fc1bc0452 not found: ID does not exist" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.538455 4888 scope.go:117] "RemoveContainer" containerID="f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea" Nov 24 00:59:35 crc kubenswrapper[4888]: E1124 00:59:35.539775 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea\": container with ID starting with f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea not found: ID does not exist" containerID="f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea" Nov 24 00:59:35 crc kubenswrapper[4888]: I1124 00:59:35.539801 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea"} err="failed to get container status \"f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea\": rpc error: code = NotFound desc = could not find container \"f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea\": container with ID starting with f731140a6886766e9cc0dad545b6f78fc8b14aa6c2e890f1476660ae5df742ea not found: ID does not exist" Nov 24 00:59:36 crc kubenswrapper[4888]: I1124 00:59:36.257730 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b06f61d-9a97-441b-808d-439551b418c5" path="/var/lib/kubelet/pods/5b06f61d-9a97-441b-808d-439551b418c5/volumes" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.166708 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n"] Nov 24 01:00:00 crc kubenswrapper[4888]: E1124 01:00:00.168083 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="extract-content" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.168109 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="extract-content" Nov 24 01:00:00 crc kubenswrapper[4888]: E1124 01:00:00.168213 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="registry-server" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.168232 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="registry-server" Nov 24 01:00:00 crc kubenswrapper[4888]: E1124 01:00:00.168253 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="extract-utilities" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.168268 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="extract-utilities" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.168692 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b06f61d-9a97-441b-808d-439551b418c5" containerName="registry-server" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.170290 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.172637 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.172707 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.180730 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n"] Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.309040 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scv72\" (UniqueName: \"kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.309151 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.309305 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.411089 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.411219 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.411363 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scv72\" (UniqueName: \"kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.413280 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.419763 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.433293 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scv72\" (UniqueName: \"kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72\") pod \"collect-profiles-29399100-7df6n\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:00 crc kubenswrapper[4888]: I1124 01:00:00.498324 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:01 crc kubenswrapper[4888]: I1124 01:00:01.077111 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n"] Nov 24 01:00:01 crc kubenswrapper[4888]: I1124 01:00:01.707646 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" event={"ID":"38d772d2-5f1c-4410-af5c-758f282527ea","Type":"ContainerStarted","Data":"3750dca87ec36b88d65df28c6d4b598a5d00baab06246896039f6a4c1cfb599a"} Nov 24 01:00:01 crc kubenswrapper[4888]: I1124 01:00:01.707941 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" event={"ID":"38d772d2-5f1c-4410-af5c-758f282527ea","Type":"ContainerStarted","Data":"98becd9e65755d79461e00e51e0d4e8f0db1a5eb49063a5ee64313d8962d1228"} Nov 24 01:00:01 crc kubenswrapper[4888]: I1124 01:00:01.729635 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" podStartSLOduration=1.7295968849999999 podStartE2EDuration="1.729596885s" podCreationTimestamp="2025-11-24 01:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:00:01.727671011 +0000 UTC m=+2104.310355065" watchObservedRunningTime="2025-11-24 01:00:01.729596885 +0000 UTC m=+2104.312280939" Nov 24 01:00:02 crc kubenswrapper[4888]: I1124 01:00:02.723879 4888 generic.go:334] "Generic (PLEG): container finished" podID="38d772d2-5f1c-4410-af5c-758f282527ea" containerID="3750dca87ec36b88d65df28c6d4b598a5d00baab06246896039f6a4c1cfb599a" exitCode=0 Nov 24 01:00:02 crc kubenswrapper[4888]: I1124 01:00:02.723953 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" event={"ID":"38d772d2-5f1c-4410-af5c-758f282527ea","Type":"ContainerDied","Data":"3750dca87ec36b88d65df28c6d4b598a5d00baab06246896039f6a4c1cfb599a"} Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.115060 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.194127 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume\") pod \"38d772d2-5f1c-4410-af5c-758f282527ea\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.194246 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scv72\" (UniqueName: \"kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72\") pod \"38d772d2-5f1c-4410-af5c-758f282527ea\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.194302 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume\") pod \"38d772d2-5f1c-4410-af5c-758f282527ea\" (UID: \"38d772d2-5f1c-4410-af5c-758f282527ea\") " Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.195118 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "38d772d2-5f1c-4410-af5c-758f282527ea" (UID: "38d772d2-5f1c-4410-af5c-758f282527ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.200299 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d772d2-5f1c-4410-af5c-758f282527ea-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.204032 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "38d772d2-5f1c-4410-af5c-758f282527ea" (UID: "38d772d2-5f1c-4410-af5c-758f282527ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.209132 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72" (OuterVolumeSpecName: "kube-api-access-scv72") pod "38d772d2-5f1c-4410-af5c-758f282527ea" (UID: "38d772d2-5f1c-4410-af5c-758f282527ea"). InnerVolumeSpecName "kube-api-access-scv72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.303197 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38d772d2-5f1c-4410-af5c-758f282527ea-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.303238 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scv72\" (UniqueName: \"kubernetes.io/projected/38d772d2-5f1c-4410-af5c-758f282527ea-kube-api-access-scv72\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.743207 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" event={"ID":"38d772d2-5f1c-4410-af5c-758f282527ea","Type":"ContainerDied","Data":"98becd9e65755d79461e00e51e0d4e8f0db1a5eb49063a5ee64313d8962d1228"} Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.743250 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98becd9e65755d79461e00e51e0d4e8f0db1a5eb49063a5ee64313d8962d1228" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.743277 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n" Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.819594 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v"] Nov 24 01:00:04 crc kubenswrapper[4888]: I1124 01:00:04.831449 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399055-tn92v"] Nov 24 01:00:06 crc kubenswrapper[4888]: I1124 01:00:06.272366 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="153d6369-b587-4da4-863c-bc8a85edde5e" path="/var/lib/kubelet/pods/153d6369-b587-4da4-863c-bc8a85edde5e/volumes" Nov 24 01:00:10 crc kubenswrapper[4888]: I1124 01:00:10.854878 4888 generic.go:334] "Generic (PLEG): container finished" podID="ca5b9705-eedc-46e0-9f41-4b070305f441" containerID="0ece25aa8b97abdb3d29e524bf4bd33c7d0ca106dfc9f77fcdc10cbcfed948aa" exitCode=0 Nov 24 01:00:10 crc kubenswrapper[4888]: I1124 01:00:10.855044 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" event={"ID":"ca5b9705-eedc-46e0-9f41-4b070305f441","Type":"ContainerDied","Data":"0ece25aa8b97abdb3d29e524bf4bd33c7d0ca106dfc9f77fcdc10cbcfed948aa"} Nov 24 01:00:11 crc kubenswrapper[4888]: I1124 01:00:11.910316 4888 scope.go:117] "RemoveContainer" containerID="c15a354f669900d9d4d784de64de4fd086f9d80c2f1a5f44ce47ef4031cb240f" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.462782 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.612863 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key\") pod \"ca5b9705-eedc-46e0-9f41-4b070305f441\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.613016 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0\") pod \"ca5b9705-eedc-46e0-9f41-4b070305f441\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.613057 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory\") pod \"ca5b9705-eedc-46e0-9f41-4b070305f441\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.613086 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6svxz\" (UniqueName: \"kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz\") pod \"ca5b9705-eedc-46e0-9f41-4b070305f441\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.613118 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle\") pod \"ca5b9705-eedc-46e0-9f41-4b070305f441\" (UID: \"ca5b9705-eedc-46e0-9f41-4b070305f441\") " Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.618594 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ca5b9705-eedc-46e0-9f41-4b070305f441" (UID: "ca5b9705-eedc-46e0-9f41-4b070305f441"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.639885 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz" (OuterVolumeSpecName: "kube-api-access-6svxz") pod "ca5b9705-eedc-46e0-9f41-4b070305f441" (UID: "ca5b9705-eedc-46e0-9f41-4b070305f441"). InnerVolumeSpecName "kube-api-access-6svxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.647924 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory" (OuterVolumeSpecName: "inventory") pod "ca5b9705-eedc-46e0-9f41-4b070305f441" (UID: "ca5b9705-eedc-46e0-9f41-4b070305f441"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.656768 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca5b9705-eedc-46e0-9f41-4b070305f441" (UID: "ca5b9705-eedc-46e0-9f41-4b070305f441"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.670127 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "ca5b9705-eedc-46e0-9f41-4b070305f441" (UID: "ca5b9705-eedc-46e0-9f41-4b070305f441"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.715371 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.715402 4888 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca5b9705-eedc-46e0-9f41-4b070305f441-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.715412 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.715421 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6svxz\" (UniqueName: \"kubernetes.io/projected/ca5b9705-eedc-46e0-9f41-4b070305f441-kube-api-access-6svxz\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.715430 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5b9705-eedc-46e0-9f41-4b070305f441-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.881188 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" event={"ID":"ca5b9705-eedc-46e0-9f41-4b070305f441","Type":"ContainerDied","Data":"7a2f12c3ab7a890ebe52003a418da8a4b3250e5abb9f77a759285dd7d992087f"} Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.881243 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a2f12c3ab7a890ebe52003a418da8a4b3250e5abb9f77a759285dd7d992087f" Nov 24 01:00:12 crc kubenswrapper[4888]: I1124 01:00:12.881338 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.030704 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj"] Nov 24 01:00:13 crc kubenswrapper[4888]: E1124 01:00:13.031132 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d772d2-5f1c-4410-af5c-758f282527ea" containerName="collect-profiles" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.031145 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d772d2-5f1c-4410-af5c-758f282527ea" containerName="collect-profiles" Nov 24 01:00:13 crc kubenswrapper[4888]: E1124 01:00:13.031183 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca5b9705-eedc-46e0-9f41-4b070305f441" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.031190 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca5b9705-eedc-46e0-9f41-4b070305f441" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.031388 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d772d2-5f1c-4410-af5c-758f282527ea" containerName="collect-profiles" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.031421 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca5b9705-eedc-46e0-9f41-4b070305f441" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.032416 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.036872 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.037096 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.037558 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.037754 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.038250 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.067113 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj"] Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.125403 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wchz9\" (UniqueName: \"kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.125531 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.125560 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.125629 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.125659 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.226901 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.226962 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.227041 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.227075 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.227109 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wchz9\" (UniqueName: \"kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.230627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.230699 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.231492 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.231870 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.246060 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wchz9\" (UniqueName: \"kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.366978 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.986776 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj"] Nov 24 01:00:13 crc kubenswrapper[4888]: W1124 01:00:13.992773 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf58fbdbc_1266_410e_b2fb_2e1aec45104e.slice/crio-e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6 WatchSource:0}: Error finding container e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6: Status 404 returned error can't find the container with id e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6 Nov 24 01:00:13 crc kubenswrapper[4888]: I1124 01:00:13.995269 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:00:14 crc kubenswrapper[4888]: I1124 01:00:14.910861 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" event={"ID":"f58fbdbc-1266-410e-b2fb-2e1aec45104e","Type":"ContainerStarted","Data":"91052a71ca06d6c79bd06daeeede9f8264d95969de631793683b516d1532de21"} Nov 24 01:00:14 crc kubenswrapper[4888]: I1124 01:00:14.911165 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" event={"ID":"f58fbdbc-1266-410e-b2fb-2e1aec45104e","Type":"ContainerStarted","Data":"e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6"} Nov 24 01:00:14 crc kubenswrapper[4888]: I1124 01:00:14.933433 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" podStartSLOduration=2.432282997 podStartE2EDuration="2.933408141s" podCreationTimestamp="2025-11-24 01:00:12 +0000 UTC" firstStartedPulling="2025-11-24 01:00:13.994973711 +0000 UTC m=+2116.577657755" lastFinishedPulling="2025-11-24 01:00:14.496098815 +0000 UTC m=+2117.078782899" observedRunningTime="2025-11-24 01:00:14.928005349 +0000 UTC m=+2117.510689443" watchObservedRunningTime="2025-11-24 01:00:14.933408141 +0000 UTC m=+2117.516092215" Nov 24 01:00:23 crc kubenswrapper[4888]: I1124 01:00:23.592330 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:00:23 crc kubenswrapper[4888]: I1124 01:00:23.593168 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:00:25 crc kubenswrapper[4888]: I1124 01:00:25.079148 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-cpbbm"] Nov 24 01:00:25 crc kubenswrapper[4888]: I1124 01:00:25.094191 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-cpbbm"] Nov 24 01:00:26 crc kubenswrapper[4888]: I1124 01:00:26.280181 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6877fe07-b619-4e4e-8b55-ed047924a852" path="/var/lib/kubelet/pods/6877fe07-b619-4e4e-8b55-ed047924a852/volumes" Nov 24 01:00:53 crc kubenswrapper[4888]: I1124 01:00:53.591753 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:00:53 crc kubenswrapper[4888]: I1124 01:00:53.592302 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:00:57 crc kubenswrapper[4888]: I1124 01:00:57.057941 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-2btp2"] Nov 24 01:00:57 crc kubenswrapper[4888]: I1124 01:00:57.069905 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-2btp2"] Nov 24 01:00:58 crc kubenswrapper[4888]: I1124 01:00:58.269399 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631c3444-214f-4fd7-9bb9-6584c5018e82" path="/var/lib/kubelet/pods/631c3444-214f-4fd7-9bb9-6584c5018e82/volumes" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.165656 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399101-qdc8s"] Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.167344 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.185148 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399101-qdc8s"] Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.305715 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.305799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.306174 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.306475 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v92hj\" (UniqueName: \"kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.409067 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.409255 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v92hj\" (UniqueName: \"kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.409494 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.409549 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.418934 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.419131 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.419685 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.442559 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v92hj\" (UniqueName: \"kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj\") pod \"keystone-cron-29399101-qdc8s\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:00 crc kubenswrapper[4888]: I1124 01:01:00.501567 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:01 crc kubenswrapper[4888]: I1124 01:01:01.004655 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399101-qdc8s"] Nov 24 01:01:01 crc kubenswrapper[4888]: I1124 01:01:01.475095 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399101-qdc8s" event={"ID":"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad","Type":"ContainerStarted","Data":"7363ff78a5dcb577cd56b0e45379729f7619c408525f319e7e8c6d09135db17a"} Nov 24 01:01:01 crc kubenswrapper[4888]: I1124 01:01:01.476661 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399101-qdc8s" event={"ID":"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad","Type":"ContainerStarted","Data":"73f045c86b74be3644bd6d0bbfc5354894a7590c1342ad69b11d20fa6fe1f51a"} Nov 24 01:01:01 crc kubenswrapper[4888]: I1124 01:01:01.495138 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399101-qdc8s" podStartSLOduration=1.495112507 podStartE2EDuration="1.495112507s" podCreationTimestamp="2025-11-24 01:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:01:01.492458043 +0000 UTC m=+2164.075142077" watchObservedRunningTime="2025-11-24 01:01:01.495112507 +0000 UTC m=+2164.077796541" Nov 24 01:01:05 crc kubenswrapper[4888]: I1124 01:01:05.515587 4888 generic.go:334] "Generic (PLEG): container finished" podID="81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" containerID="7363ff78a5dcb577cd56b0e45379729f7619c408525f319e7e8c6d09135db17a" exitCode=0 Nov 24 01:01:05 crc kubenswrapper[4888]: I1124 01:01:05.515675 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399101-qdc8s" event={"ID":"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad","Type":"ContainerDied","Data":"7363ff78a5dcb577cd56b0e45379729f7619c408525f319e7e8c6d09135db17a"} Nov 24 01:01:06 crc kubenswrapper[4888]: I1124 01:01:06.955103 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.084036 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data\") pod \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.084142 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys\") pod \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.084395 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle\") pod \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.084452 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v92hj\" (UniqueName: \"kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj\") pod \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\" (UID: \"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad\") " Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.094707 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj" (OuterVolumeSpecName: "kube-api-access-v92hj") pod "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" (UID: "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad"). InnerVolumeSpecName "kube-api-access-v92hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.096901 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" (UID: "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.117698 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" (UID: "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.174335 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data" (OuterVolumeSpecName: "config-data") pod "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" (UID: "81ab4fd4-2f1c-4023-b7a1-528e59bb56ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.200566 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.200606 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v92hj\" (UniqueName: \"kubernetes.io/projected/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-kube-api-access-v92hj\") on node \"crc\" DevicePath \"\"" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.200623 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.200634 4888 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81ab4fd4-2f1c-4023-b7a1-528e59bb56ad-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.545473 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399101-qdc8s" event={"ID":"81ab4fd4-2f1c-4023-b7a1-528e59bb56ad","Type":"ContainerDied","Data":"73f045c86b74be3644bd6d0bbfc5354894a7590c1342ad69b11d20fa6fe1f51a"} Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.545902 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f045c86b74be3644bd6d0bbfc5354894a7590c1342ad69b11d20fa6fe1f51a" Nov 24 01:01:07 crc kubenswrapper[4888]: I1124 01:01:07.545552 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399101-qdc8s" Nov 24 01:01:12 crc kubenswrapper[4888]: I1124 01:01:12.002501 4888 scope.go:117] "RemoveContainer" containerID="109eaa690a489aed893a747b15ff79dabe2790e0055a5a5f4d43e78931550e51" Nov 24 01:01:12 crc kubenswrapper[4888]: I1124 01:01:12.055625 4888 scope.go:117] "RemoveContainer" containerID="6cadd228fdd7a0a141c0eb92bd0de6b2524d3a9f1b97fad58422fc078f1eb641" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.592360 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.593136 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.593207 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.594519 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.594620 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" gracePeriod=600 Nov 24 01:01:23 crc kubenswrapper[4888]: E1124 01:01:23.727510 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.750219 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" exitCode=0 Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.750285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1"} Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.750372 4888 scope.go:117] "RemoveContainer" containerID="09d8ca21a91ea9afbea6d41c7887b6d763f9247c068a35be1da5edc793178e63" Nov 24 01:01:23 crc kubenswrapper[4888]: I1124 01:01:23.751254 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:01:23 crc kubenswrapper[4888]: E1124 01:01:23.751700 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:01:36 crc kubenswrapper[4888]: I1124 01:01:36.245765 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:01:36 crc kubenswrapper[4888]: E1124 01:01:36.246633 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:01:51 crc kubenswrapper[4888]: I1124 01:01:51.246530 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:01:51 crc kubenswrapper[4888]: E1124 01:01:51.247528 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:02:02 crc kubenswrapper[4888]: I1124 01:02:02.246665 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:02:02 crc kubenswrapper[4888]: E1124 01:02:02.247825 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:02:14 crc kubenswrapper[4888]: I1124 01:02:14.246447 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:02:14 crc kubenswrapper[4888]: E1124 01:02:14.247719 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:02:26 crc kubenswrapper[4888]: I1124 01:02:26.249566 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:02:26 crc kubenswrapper[4888]: E1124 01:02:26.250437 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:02:41 crc kubenswrapper[4888]: I1124 01:02:41.246117 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:02:41 crc kubenswrapper[4888]: E1124 01:02:41.247121 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:02:55 crc kubenswrapper[4888]: I1124 01:02:55.245676 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:02:55 crc kubenswrapper[4888]: E1124 01:02:55.246927 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:03:10 crc kubenswrapper[4888]: I1124 01:03:10.245886 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:03:10 crc kubenswrapper[4888]: E1124 01:03:10.247963 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:03:25 crc kubenswrapper[4888]: I1124 01:03:25.244920 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:03:25 crc kubenswrapper[4888]: E1124 01:03:25.245760 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:03:37 crc kubenswrapper[4888]: I1124 01:03:37.606906 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:03:37 crc kubenswrapper[4888]: E1124 01:03:37.608833 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:03:51 crc kubenswrapper[4888]: I1124 01:03:51.246618 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:03:51 crc kubenswrapper[4888]: E1124 01:03:51.247498 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:04:05 crc kubenswrapper[4888]: I1124 01:04:05.246226 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:04:05 crc kubenswrapper[4888]: E1124 01:04:05.247494 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:04:17 crc kubenswrapper[4888]: I1124 01:04:17.245499 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:04:17 crc kubenswrapper[4888]: E1124 01:04:17.246582 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:04:32 crc kubenswrapper[4888]: I1124 01:04:32.246162 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:04:32 crc kubenswrapper[4888]: E1124 01:04:32.247512 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:04:44 crc kubenswrapper[4888]: I1124 01:04:44.246387 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:04:44 crc kubenswrapper[4888]: E1124 01:04:44.248044 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:04:46 crc kubenswrapper[4888]: I1124 01:04:46.563644 4888 generic.go:334] "Generic (PLEG): container finished" podID="f58fbdbc-1266-410e-b2fb-2e1aec45104e" containerID="91052a71ca06d6c79bd06daeeede9f8264d95969de631793683b516d1532de21" exitCode=0 Nov 24 01:04:46 crc kubenswrapper[4888]: I1124 01:04:46.563729 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" event={"ID":"f58fbdbc-1266-410e-b2fb-2e1aec45104e","Type":"ContainerDied","Data":"91052a71ca06d6c79bd06daeeede9f8264d95969de631793683b516d1532de21"} Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.123893 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.300051 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory\") pod \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.300137 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle\") pod \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.300175 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0\") pod \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.300485 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key\") pod \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.300587 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wchz9\" (UniqueName: \"kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9\") pod \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\" (UID: \"f58fbdbc-1266-410e-b2fb-2e1aec45104e\") " Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.307366 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9" (OuterVolumeSpecName: "kube-api-access-wchz9") pod "f58fbdbc-1266-410e-b2fb-2e1aec45104e" (UID: "f58fbdbc-1266-410e-b2fb-2e1aec45104e"). InnerVolumeSpecName "kube-api-access-wchz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.318174 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f58fbdbc-1266-410e-b2fb-2e1aec45104e" (UID: "f58fbdbc-1266-410e-b2fb-2e1aec45104e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.341598 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f58fbdbc-1266-410e-b2fb-2e1aec45104e" (UID: "f58fbdbc-1266-410e-b2fb-2e1aec45104e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.356012 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory" (OuterVolumeSpecName: "inventory") pod "f58fbdbc-1266-410e-b2fb-2e1aec45104e" (UID: "f58fbdbc-1266-410e-b2fb-2e1aec45104e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.358911 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f58fbdbc-1266-410e-b2fb-2e1aec45104e" (UID: "f58fbdbc-1266-410e-b2fb-2e1aec45104e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.404762 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.404850 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.404876 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.404893 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f58fbdbc-1266-410e-b2fb-2e1aec45104e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.404980 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wchz9\" (UniqueName: \"kubernetes.io/projected/f58fbdbc-1266-410e-b2fb-2e1aec45104e-kube-api-access-wchz9\") on node \"crc\" DevicePath \"\"" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.598647 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" event={"ID":"f58fbdbc-1266-410e-b2fb-2e1aec45104e","Type":"ContainerDied","Data":"e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6"} Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.598712 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f3e3c581a8522b3d15e64b7139e9be3d31760b12ee2e34a36cb2f6bd8824d6" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.598767 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.720334 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl"] Nov 24 01:04:48 crc kubenswrapper[4888]: E1124 01:04:48.720732 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" containerName="keystone-cron" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.720748 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" containerName="keystone-cron" Nov 24 01:04:48 crc kubenswrapper[4888]: E1124 01:04:48.720763 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58fbdbc-1266-410e-b2fb-2e1aec45104e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.720770 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58fbdbc-1266-410e-b2fb-2e1aec45104e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.721013 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f58fbdbc-1266-410e-b2fb-2e1aec45104e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.721031 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ab4fd4-2f1c-4023-b7a1-528e59bb56ad" containerName="keystone-cron" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.722028 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.724866 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.725292 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.725319 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.725466 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.725999 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.731568 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl"] Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.917933 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.917996 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.918086 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.918107 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp5vx\" (UniqueName: \"kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.918139 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.918158 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:48 crc kubenswrapper[4888]: I1124 01:04:48.918195 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020540 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020616 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020730 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020752 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp5vx\" (UniqueName: \"kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020801 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020851 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.020903 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.025875 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.026353 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.026356 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.027529 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.028355 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.028432 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.047271 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp5vx\" (UniqueName: \"kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jdndl\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.341063 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:04:49 crc kubenswrapper[4888]: I1124 01:04:49.941310 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl"] Nov 24 01:04:50 crc kubenswrapper[4888]: I1124 01:04:50.637017 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" event={"ID":"371f2fbd-a533-4ac3-85e8-68b5c3dccc25","Type":"ContainerStarted","Data":"9aa5b1ae55c5bfe75937da9cdc3ddf1f48ff74ec9139b0b4ba9a9dd7c0e5ba47"} Nov 24 01:04:51 crc kubenswrapper[4888]: I1124 01:04:51.647856 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" event={"ID":"371f2fbd-a533-4ac3-85e8-68b5c3dccc25","Type":"ContainerStarted","Data":"37173908abef59b4695d933068c59d24092cb8036e13dc4b8a72d6806c7daef6"} Nov 24 01:04:51 crc kubenswrapper[4888]: I1124 01:04:51.670677 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" podStartSLOduration=2.772401973 podStartE2EDuration="3.670661705s" podCreationTimestamp="2025-11-24 01:04:48 +0000 UTC" firstStartedPulling="2025-11-24 01:04:49.960449537 +0000 UTC m=+2392.543133581" lastFinishedPulling="2025-11-24 01:04:50.858709269 +0000 UTC m=+2393.441393313" observedRunningTime="2025-11-24 01:04:51.665984084 +0000 UTC m=+2394.248668158" watchObservedRunningTime="2025-11-24 01:04:51.670661705 +0000 UTC m=+2394.253345749" Nov 24 01:04:57 crc kubenswrapper[4888]: I1124 01:04:57.245394 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:04:57 crc kubenswrapper[4888]: E1124 01:04:57.246220 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:05:08 crc kubenswrapper[4888]: I1124 01:05:08.263011 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:05:08 crc kubenswrapper[4888]: E1124 01:05:08.264163 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:05:19 crc kubenswrapper[4888]: I1124 01:05:19.245692 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:05:19 crc kubenswrapper[4888]: E1124 01:05:19.246906 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:05:31 crc kubenswrapper[4888]: I1124 01:05:31.245781 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:05:31 crc kubenswrapper[4888]: E1124 01:05:31.247019 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:05:43 crc kubenswrapper[4888]: I1124 01:05:43.245361 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:05:43 crc kubenswrapper[4888]: E1124 01:05:43.246322 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:05:57 crc kubenswrapper[4888]: I1124 01:05:57.245180 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:05:57 crc kubenswrapper[4888]: E1124 01:05:57.246961 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.600140 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.603042 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.629474 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.688321 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.688414 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxfvq\" (UniqueName: \"kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.688683 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.790770 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxfvq\" (UniqueName: \"kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.790993 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.791161 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.791558 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.791602 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.813331 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxfvq\" (UniqueName: \"kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq\") pod \"community-operators-nskfm\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:04 crc kubenswrapper[4888]: I1124 01:06:04.935662 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:05 crc kubenswrapper[4888]: I1124 01:06:05.484120 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:05 crc kubenswrapper[4888]: I1124 01:06:05.580091 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerStarted","Data":"0f137fca894062893a7054e002cfe7d14dbc546976e8bf7106adfc0bea718c2a"} Nov 24 01:06:06 crc kubenswrapper[4888]: I1124 01:06:06.590994 4888 generic.go:334] "Generic (PLEG): container finished" podID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerID="6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f" exitCode=0 Nov 24 01:06:06 crc kubenswrapper[4888]: I1124 01:06:06.591040 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerDied","Data":"6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f"} Nov 24 01:06:06 crc kubenswrapper[4888]: I1124 01:06:06.593779 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:06:08 crc kubenswrapper[4888]: I1124 01:06:08.617281 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerStarted","Data":"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9"} Nov 24 01:06:09 crc kubenswrapper[4888]: I1124 01:06:09.630343 4888 generic.go:334] "Generic (PLEG): container finished" podID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerID="278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9" exitCode=0 Nov 24 01:06:09 crc kubenswrapper[4888]: I1124 01:06:09.630402 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerDied","Data":"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9"} Nov 24 01:06:10 crc kubenswrapper[4888]: I1124 01:06:10.247680 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:06:10 crc kubenswrapper[4888]: E1124 01:06:10.248390 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:06:10 crc kubenswrapper[4888]: I1124 01:06:10.644370 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerStarted","Data":"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5"} Nov 24 01:06:10 crc kubenswrapper[4888]: I1124 01:06:10.668329 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nskfm" podStartSLOduration=3.216535439 podStartE2EDuration="6.668306841s" podCreationTimestamp="2025-11-24 01:06:04 +0000 UTC" firstStartedPulling="2025-11-24 01:06:06.593521913 +0000 UTC m=+2469.176205967" lastFinishedPulling="2025-11-24 01:06:10.045293325 +0000 UTC m=+2472.627977369" observedRunningTime="2025-11-24 01:06:10.664908615 +0000 UTC m=+2473.247592699" watchObservedRunningTime="2025-11-24 01:06:10.668306841 +0000 UTC m=+2473.250990885" Nov 24 01:06:14 crc kubenswrapper[4888]: I1124 01:06:14.936784 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:14 crc kubenswrapper[4888]: I1124 01:06:14.937467 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:15 crc kubenswrapper[4888]: I1124 01:06:15.011078 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:15 crc kubenswrapper[4888]: I1124 01:06:15.809746 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:15 crc kubenswrapper[4888]: I1124 01:06:15.880390 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:17 crc kubenswrapper[4888]: I1124 01:06:17.722944 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nskfm" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="registry-server" containerID="cri-o://2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5" gracePeriod=2 Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.228269 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.306356 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxfvq\" (UniqueName: \"kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq\") pod \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.306755 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities\") pod \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.306860 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content\") pod \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\" (UID: \"06b811f6-3c40-4abc-b534-eb0c872a3ce0\") " Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.308034 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities" (OuterVolumeSpecName: "utilities") pod "06b811f6-3c40-4abc-b534-eb0c872a3ce0" (UID: "06b811f6-3c40-4abc-b534-eb0c872a3ce0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.313284 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq" (OuterVolumeSpecName: "kube-api-access-bxfvq") pod "06b811f6-3c40-4abc-b534-eb0c872a3ce0" (UID: "06b811f6-3c40-4abc-b534-eb0c872a3ce0"). InnerVolumeSpecName "kube-api-access-bxfvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.370794 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06b811f6-3c40-4abc-b534-eb0c872a3ce0" (UID: "06b811f6-3c40-4abc-b534-eb0c872a3ce0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.409078 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.409143 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxfvq\" (UniqueName: \"kubernetes.io/projected/06b811f6-3c40-4abc-b534-eb0c872a3ce0-kube-api-access-bxfvq\") on node \"crc\" DevicePath \"\"" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.409159 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b811f6-3c40-4abc-b534-eb0c872a3ce0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.736758 4888 generic.go:334] "Generic (PLEG): container finished" podID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerID="2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5" exitCode=0 Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.736840 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerDied","Data":"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5"} Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.736880 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nskfm" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.737110 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nskfm" event={"ID":"06b811f6-3c40-4abc-b534-eb0c872a3ce0","Type":"ContainerDied","Data":"0f137fca894062893a7054e002cfe7d14dbc546976e8bf7106adfc0bea718c2a"} Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.737124 4888 scope.go:117] "RemoveContainer" containerID="2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.767874 4888 scope.go:117] "RemoveContainer" containerID="278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.805577 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.819134 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nskfm"] Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.821275 4888 scope.go:117] "RemoveContainer" containerID="6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.870898 4888 scope.go:117] "RemoveContainer" containerID="2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5" Nov 24 01:06:18 crc kubenswrapper[4888]: E1124 01:06:18.871452 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5\": container with ID starting with 2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5 not found: ID does not exist" containerID="2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.871481 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5"} err="failed to get container status \"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5\": rpc error: code = NotFound desc = could not find container \"2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5\": container with ID starting with 2614e3f0ff7fd871fdbce4db4bd95731a057748659708155bf0eb59e048e82d5 not found: ID does not exist" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.871501 4888 scope.go:117] "RemoveContainer" containerID="278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9" Nov 24 01:06:18 crc kubenswrapper[4888]: E1124 01:06:18.871912 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9\": container with ID starting with 278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9 not found: ID does not exist" containerID="278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.871972 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9"} err="failed to get container status \"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9\": rpc error: code = NotFound desc = could not find container \"278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9\": container with ID starting with 278decdba1033f98a4850bb7db1eed510778930de627d22a8913582231210de9 not found: ID does not exist" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.872006 4888 scope.go:117] "RemoveContainer" containerID="6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f" Nov 24 01:06:18 crc kubenswrapper[4888]: E1124 01:06:18.872335 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f\": container with ID starting with 6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f not found: ID does not exist" containerID="6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f" Nov 24 01:06:18 crc kubenswrapper[4888]: I1124 01:06:18.872365 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f"} err="failed to get container status \"6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f\": rpc error: code = NotFound desc = could not find container \"6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f\": container with ID starting with 6483e81028a4f80c5bce3aec3ab2a3834102bfa54bb243341699a0382c87465f not found: ID does not exist" Nov 24 01:06:20 crc kubenswrapper[4888]: I1124 01:06:20.268339 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" path="/var/lib/kubelet/pods/06b811f6-3c40-4abc-b534-eb0c872a3ce0/volumes" Nov 24 01:06:24 crc kubenswrapper[4888]: I1124 01:06:24.244950 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:06:24 crc kubenswrapper[4888]: I1124 01:06:24.835327 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4"} Nov 24 01:07:34 crc kubenswrapper[4888]: I1124 01:07:34.751631 4888 generic.go:334] "Generic (PLEG): container finished" podID="371f2fbd-a533-4ac3-85e8-68b5c3dccc25" containerID="37173908abef59b4695d933068c59d24092cb8036e13dc4b8a72d6806c7daef6" exitCode=0 Nov 24 01:07:34 crc kubenswrapper[4888]: I1124 01:07:34.751680 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" event={"ID":"371f2fbd-a533-4ac3-85e8-68b5c3dccc25","Type":"ContainerDied","Data":"37173908abef59b4695d933068c59d24092cb8036e13dc4b8a72d6806c7daef6"} Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.358906 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.553874 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.553977 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.554067 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp5vx\" (UniqueName: \"kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.554139 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.554216 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.554247 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.554290 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1\") pod \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\" (UID: \"371f2fbd-a533-4ac3-85e8-68b5c3dccc25\") " Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.561488 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx" (OuterVolumeSpecName: "kube-api-access-vp5vx") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "kube-api-access-vp5vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.561605 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.586415 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.606032 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.607640 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.607729 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory" (OuterVolumeSpecName: "inventory") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.608263 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "371f2fbd-a533-4ac3-85e8-68b5c3dccc25" (UID: "371f2fbd-a533-4ac3-85e8-68b5c3dccc25"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657584 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657628 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp5vx\" (UniqueName: \"kubernetes.io/projected/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-kube-api-access-vp5vx\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657646 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657658 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657670 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657681 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.657693 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/371f2fbd-a533-4ac3-85e8-68b5c3dccc25-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.780805 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" event={"ID":"371f2fbd-a533-4ac3-85e8-68b5c3dccc25","Type":"ContainerDied","Data":"9aa5b1ae55c5bfe75937da9cdc3ddf1f48ff74ec9139b0b4ba9a9dd7c0e5ba47"} Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.781285 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aa5b1ae55c5bfe75937da9cdc3ddf1f48ff74ec9139b0b4ba9a9dd7c0e5ba47" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.781237 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.933014 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm"] Nov 24 01:07:36 crc kubenswrapper[4888]: E1124 01:07:36.933470 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="extract-content" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.933491 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="extract-content" Nov 24 01:07:36 crc kubenswrapper[4888]: E1124 01:07:36.933513 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="registry-server" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.933522 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="registry-server" Nov 24 01:07:36 crc kubenswrapper[4888]: E1124 01:07:36.933541 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="extract-utilities" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.933550 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="extract-utilities" Nov 24 01:07:36 crc kubenswrapper[4888]: E1124 01:07:36.933561 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371f2fbd-a533-4ac3-85e8-68b5c3dccc25" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.933570 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="371f2fbd-a533-4ac3-85e8-68b5c3dccc25" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.934009 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="371f2fbd-a533-4ac3-85e8-68b5c3dccc25" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.934035 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b811f6-3c40-4abc-b534-eb0c872a3ce0" containerName="registry-server" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.934870 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.937584 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.937717 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.938008 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.938071 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.939440 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:07:36 crc kubenswrapper[4888]: I1124 01:07:36.948903 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm"] Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065062 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065185 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065246 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065382 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zslvl\" (UniqueName: \"kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065544 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065623 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.065669 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.167991 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168063 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168128 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168191 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168246 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168295 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zslvl\" (UniqueName: \"kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.168395 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.173774 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.174919 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.177983 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.179246 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.181346 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.186608 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.190894 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zslvl\" (UniqueName: \"kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.257986 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:07:37 crc kubenswrapper[4888]: I1124 01:07:37.856359 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm"] Nov 24 01:07:38 crc kubenswrapper[4888]: I1124 01:07:38.821865 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" event={"ID":"fdba4390-e81c-410b-9718-3d87e5ca4d82","Type":"ContainerStarted","Data":"1d9dc355926e87e804a695da62eb6262077008a8e036b025803bf41c728d8215"} Nov 24 01:07:39 crc kubenswrapper[4888]: I1124 01:07:39.839294 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" event={"ID":"fdba4390-e81c-410b-9718-3d87e5ca4d82","Type":"ContainerStarted","Data":"ba0fd62114ef12169c9e8244b3c6d625df5019f48d7cffb5aab7cd8629eed080"} Nov 24 01:07:39 crc kubenswrapper[4888]: I1124 01:07:39.871896 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" podStartSLOduration=3.305014352 podStartE2EDuration="3.871871078s" podCreationTimestamp="2025-11-24 01:07:36 +0000 UTC" firstStartedPulling="2025-11-24 01:07:37.872108683 +0000 UTC m=+2560.454792717" lastFinishedPulling="2025-11-24 01:07:38.438965399 +0000 UTC m=+2561.021649443" observedRunningTime="2025-11-24 01:07:39.860407536 +0000 UTC m=+2562.443091600" watchObservedRunningTime="2025-11-24 01:07:39.871871078 +0000 UTC m=+2562.454555143" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.098498 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.104748 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.120731 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.268974 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.269304 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56pz5\" (UniqueName: \"kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.269406 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.371108 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.371274 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.371306 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56pz5\" (UniqueName: \"kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.371619 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.372060 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.393954 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56pz5\" (UniqueName: \"kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5\") pod \"certified-operators-v2pff\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.444907 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:11 crc kubenswrapper[4888]: I1124 01:08:11.970001 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:12 crc kubenswrapper[4888]: I1124 01:08:12.189362 4888 generic.go:334] "Generic (PLEG): container finished" podID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerID="67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2" exitCode=0 Nov 24 01:08:12 crc kubenswrapper[4888]: I1124 01:08:12.189425 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerDied","Data":"67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2"} Nov 24 01:08:12 crc kubenswrapper[4888]: I1124 01:08:12.189611 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerStarted","Data":"c3e7e5f0f3f79389a58b87b708be5e1e09c25f4993944389426e14baa2dac1bc"} Nov 24 01:08:13 crc kubenswrapper[4888]: I1124 01:08:13.201250 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerStarted","Data":"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a"} Nov 24 01:08:15 crc kubenswrapper[4888]: I1124 01:08:15.235734 4888 generic.go:334] "Generic (PLEG): container finished" podID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerID="c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a" exitCode=0 Nov 24 01:08:15 crc kubenswrapper[4888]: I1124 01:08:15.235850 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerDied","Data":"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a"} Nov 24 01:08:16 crc kubenswrapper[4888]: I1124 01:08:16.257238 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerStarted","Data":"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415"} Nov 24 01:08:16 crc kubenswrapper[4888]: I1124 01:08:16.283379 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v2pff" podStartSLOduration=1.85277111 podStartE2EDuration="5.283363165s" podCreationTimestamp="2025-11-24 01:08:11 +0000 UTC" firstStartedPulling="2025-11-24 01:08:12.190952152 +0000 UTC m=+2594.773636206" lastFinishedPulling="2025-11-24 01:08:15.621544177 +0000 UTC m=+2598.204228261" observedRunningTime="2025-11-24 01:08:16.277687576 +0000 UTC m=+2598.860371620" watchObservedRunningTime="2025-11-24 01:08:16.283363165 +0000 UTC m=+2598.866047209" Nov 24 01:08:21 crc kubenswrapper[4888]: I1124 01:08:21.445025 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:21 crc kubenswrapper[4888]: I1124 01:08:21.445695 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:21 crc kubenswrapper[4888]: I1124 01:08:21.524376 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:22 crc kubenswrapper[4888]: I1124 01:08:22.407266 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:22 crc kubenswrapper[4888]: I1124 01:08:22.482725 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.343749 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v2pff" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="registry-server" containerID="cri-o://07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415" gracePeriod=2 Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.901601 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.997440 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities\") pod \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.997517 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content\") pod \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.997677 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56pz5\" (UniqueName: \"kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5\") pod \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\" (UID: \"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d\") " Nov 24 01:08:24 crc kubenswrapper[4888]: I1124 01:08:24.999686 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities" (OuterVolumeSpecName: "utilities") pod "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" (UID: "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.005119 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5" (OuterVolumeSpecName: "kube-api-access-56pz5") pod "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" (UID: "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d"). InnerVolumeSpecName "kube-api-access-56pz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.052117 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" (UID: "b1b43d7b-ea3f-49c4-9457-0432b5c2f56d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.100553 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.100611 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.100633 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56pz5\" (UniqueName: \"kubernetes.io/projected/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d-kube-api-access-56pz5\") on node \"crc\" DevicePath \"\"" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.356465 4888 generic.go:334] "Generic (PLEG): container finished" podID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerID="07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415" exitCode=0 Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.356512 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerDied","Data":"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415"} Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.356538 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2pff" event={"ID":"b1b43d7b-ea3f-49c4-9457-0432b5c2f56d","Type":"ContainerDied","Data":"c3e7e5f0f3f79389a58b87b708be5e1e09c25f4993944389426e14baa2dac1bc"} Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.356541 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2pff" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.356565 4888 scope.go:117] "RemoveContainer" containerID="07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.378302 4888 scope.go:117] "RemoveContainer" containerID="c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.426900 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.436979 4888 scope.go:117] "RemoveContainer" containerID="67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.461698 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v2pff"] Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.488942 4888 scope.go:117] "RemoveContainer" containerID="07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415" Nov 24 01:08:25 crc kubenswrapper[4888]: E1124 01:08:25.489541 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415\": container with ID starting with 07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415 not found: ID does not exist" containerID="07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.489572 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415"} err="failed to get container status \"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415\": rpc error: code = NotFound desc = could not find container \"07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415\": container with ID starting with 07e487d5177fcd7ca24c920e053189cb5d20aca58a64285ae2fb32e34d99b415 not found: ID does not exist" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.489592 4888 scope.go:117] "RemoveContainer" containerID="c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a" Nov 24 01:08:25 crc kubenswrapper[4888]: E1124 01:08:25.490919 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a\": container with ID starting with c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a not found: ID does not exist" containerID="c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.490944 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a"} err="failed to get container status \"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a\": rpc error: code = NotFound desc = could not find container \"c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a\": container with ID starting with c8fc15aaf7092ea1deda08db1b8876cbb072e94c143c6bcbacad4681b5554b1a not found: ID does not exist" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.490958 4888 scope.go:117] "RemoveContainer" containerID="67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2" Nov 24 01:08:25 crc kubenswrapper[4888]: E1124 01:08:25.494155 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2\": container with ID starting with 67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2 not found: ID does not exist" containerID="67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2" Nov 24 01:08:25 crc kubenswrapper[4888]: I1124 01:08:25.494200 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2"} err="failed to get container status \"67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2\": rpc error: code = NotFound desc = could not find container \"67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2\": container with ID starting with 67f21f3772ad9cb96e0312b3f5053fb8ec78bfe83fc86a64b5abb909c3cf90e2 not found: ID does not exist" Nov 24 01:08:26 crc kubenswrapper[4888]: I1124 01:08:26.264411 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" path="/var/lib/kubelet/pods/b1b43d7b-ea3f-49c4-9457-0432b5c2f56d/volumes" Nov 24 01:08:53 crc kubenswrapper[4888]: I1124 01:08:53.591432 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:08:53 crc kubenswrapper[4888]: I1124 01:08:53.592036 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.124256 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:05 crc kubenswrapper[4888]: E1124 01:09:05.126355 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="extract-content" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.126379 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="extract-content" Nov 24 01:09:05 crc kubenswrapper[4888]: E1124 01:09:05.126398 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="registry-server" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.126404 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="registry-server" Nov 24 01:09:05 crc kubenswrapper[4888]: E1124 01:09:05.126434 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="extract-utilities" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.126444 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="extract-utilities" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.126675 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b43d7b-ea3f-49c4-9457-0432b5c2f56d" containerName="registry-server" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.128595 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.139844 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.206635 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp5mc\" (UniqueName: \"kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.207082 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.207143 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.309251 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.309296 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.309412 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp5mc\" (UniqueName: \"kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.309899 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.310102 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.331480 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp5mc\" (UniqueName: \"kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc\") pod \"redhat-marketplace-j7pfd\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.463123 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:05 crc kubenswrapper[4888]: I1124 01:09:05.939218 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:05 crc kubenswrapper[4888]: W1124 01:09:05.947274 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefa86846_b07e_45d9_b7db_eb4a088403b0.slice/crio-7aa2a35ad6c0d458085befcc7484e3de054286ffc0b3cecf82c13e663bfd0ec1 WatchSource:0}: Error finding container 7aa2a35ad6c0d458085befcc7484e3de054286ffc0b3cecf82c13e663bfd0ec1: Status 404 returned error can't find the container with id 7aa2a35ad6c0d458085befcc7484e3de054286ffc0b3cecf82c13e663bfd0ec1 Nov 24 01:09:06 crc kubenswrapper[4888]: I1124 01:09:06.884403 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerDied","Data":"ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b"} Nov 24 01:09:06 crc kubenswrapper[4888]: I1124 01:09:06.884221 4888 generic.go:334] "Generic (PLEG): container finished" podID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerID="ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b" exitCode=0 Nov 24 01:09:06 crc kubenswrapper[4888]: I1124 01:09:06.884982 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerStarted","Data":"7aa2a35ad6c0d458085befcc7484e3de054286ffc0b3cecf82c13e663bfd0ec1"} Nov 24 01:09:08 crc kubenswrapper[4888]: I1124 01:09:08.914844 4888 generic.go:334] "Generic (PLEG): container finished" podID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerID="245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c" exitCode=0 Nov 24 01:09:08 crc kubenswrapper[4888]: I1124 01:09:08.914916 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerDied","Data":"245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c"} Nov 24 01:09:09 crc kubenswrapper[4888]: I1124 01:09:09.931214 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerStarted","Data":"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513"} Nov 24 01:09:09 crc kubenswrapper[4888]: I1124 01:09:09.957780 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j7pfd" podStartSLOduration=2.506774514 podStartE2EDuration="4.957757692s" podCreationTimestamp="2025-11-24 01:09:05 +0000 UTC" firstStartedPulling="2025-11-24 01:09:06.893027059 +0000 UTC m=+2649.475711143" lastFinishedPulling="2025-11-24 01:09:09.344010247 +0000 UTC m=+2651.926694321" observedRunningTime="2025-11-24 01:09:09.948535853 +0000 UTC m=+2652.531219907" watchObservedRunningTime="2025-11-24 01:09:09.957757692 +0000 UTC m=+2652.540441736" Nov 24 01:09:15 crc kubenswrapper[4888]: I1124 01:09:15.464604 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:15 crc kubenswrapper[4888]: I1124 01:09:15.465256 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:15 crc kubenswrapper[4888]: I1124 01:09:15.557152 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:16 crc kubenswrapper[4888]: I1124 01:09:16.067387 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:17 crc kubenswrapper[4888]: I1124 01:09:17.105877 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.040173 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j7pfd" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="registry-server" containerID="cri-o://5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513" gracePeriod=2 Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.597496 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.736074 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp5mc\" (UniqueName: \"kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc\") pod \"efa86846-b07e-45d9-b7db-eb4a088403b0\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.736570 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities\") pod \"efa86846-b07e-45d9-b7db-eb4a088403b0\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.736627 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content\") pod \"efa86846-b07e-45d9-b7db-eb4a088403b0\" (UID: \"efa86846-b07e-45d9-b7db-eb4a088403b0\") " Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.737488 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities" (OuterVolumeSpecName: "utilities") pod "efa86846-b07e-45d9-b7db-eb4a088403b0" (UID: "efa86846-b07e-45d9-b7db-eb4a088403b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.747732 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc" (OuterVolumeSpecName: "kube-api-access-wp5mc") pod "efa86846-b07e-45d9-b7db-eb4a088403b0" (UID: "efa86846-b07e-45d9-b7db-eb4a088403b0"). InnerVolumeSpecName "kube-api-access-wp5mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.752856 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efa86846-b07e-45d9-b7db-eb4a088403b0" (UID: "efa86846-b07e-45d9-b7db-eb4a088403b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.838559 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp5mc\" (UniqueName: \"kubernetes.io/projected/efa86846-b07e-45d9-b7db-eb4a088403b0-kube-api-access-wp5mc\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.838590 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:18 crc kubenswrapper[4888]: I1124 01:09:18.838599 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa86846-b07e-45d9-b7db-eb4a088403b0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.050737 4888 generic.go:334] "Generic (PLEG): container finished" podID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerID="5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513" exitCode=0 Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.050788 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerDied","Data":"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513"} Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.050838 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pfd" event={"ID":"efa86846-b07e-45d9-b7db-eb4a088403b0","Type":"ContainerDied","Data":"7aa2a35ad6c0d458085befcc7484e3de054286ffc0b3cecf82c13e663bfd0ec1"} Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.050858 4888 scope.go:117] "RemoveContainer" containerID="5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.051005 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pfd" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.082182 4888 scope.go:117] "RemoveContainer" containerID="245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.088546 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.097445 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pfd"] Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.104965 4888 scope.go:117] "RemoveContainer" containerID="ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.149275 4888 scope.go:117] "RemoveContainer" containerID="5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513" Nov 24 01:09:19 crc kubenswrapper[4888]: E1124 01:09:19.149627 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513\": container with ID starting with 5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513 not found: ID does not exist" containerID="5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.149676 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513"} err="failed to get container status \"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513\": rpc error: code = NotFound desc = could not find container \"5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513\": container with ID starting with 5986bcf670b266cf4ffe70dcbe507e6256db820a2ccafe58862ea4a55909a513 not found: ID does not exist" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.149708 4888 scope.go:117] "RemoveContainer" containerID="245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c" Nov 24 01:09:19 crc kubenswrapper[4888]: E1124 01:09:19.149917 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c\": container with ID starting with 245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c not found: ID does not exist" containerID="245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.149945 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c"} err="failed to get container status \"245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c\": rpc error: code = NotFound desc = could not find container \"245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c\": container with ID starting with 245dcf9ca49febb5740858e198077956f50b725c03d02d8410d947e1d6af319c not found: ID does not exist" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.149999 4888 scope.go:117] "RemoveContainer" containerID="ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b" Nov 24 01:09:19 crc kubenswrapper[4888]: E1124 01:09:19.150187 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b\": container with ID starting with ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b not found: ID does not exist" containerID="ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b" Nov 24 01:09:19 crc kubenswrapper[4888]: I1124 01:09:19.150219 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b"} err="failed to get container status \"ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b\": rpc error: code = NotFound desc = could not find container \"ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b\": container with ID starting with ae50a5a92f2e3e352a63678b22a53fec4bd845460b2d2bf00a5cde7daa36455b not found: ID does not exist" Nov 24 01:09:20 crc kubenswrapper[4888]: I1124 01:09:20.277213 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" path="/var/lib/kubelet/pods/efa86846-b07e-45d9-b7db-eb4a088403b0/volumes" Nov 24 01:09:23 crc kubenswrapper[4888]: I1124 01:09:23.591588 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:09:23 crc kubenswrapper[4888]: I1124 01:09:23.592287 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:09:53 crc kubenswrapper[4888]: I1124 01:09:53.591228 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:09:53 crc kubenswrapper[4888]: I1124 01:09:53.592218 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:09:53 crc kubenswrapper[4888]: I1124 01:09:53.592287 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:09:53 crc kubenswrapper[4888]: I1124 01:09:53.593058 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:09:53 crc kubenswrapper[4888]: I1124 01:09:53.593146 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4" gracePeriod=600 Nov 24 01:09:54 crc kubenswrapper[4888]: I1124 01:09:54.536844 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4" exitCode=0 Nov 24 01:09:54 crc kubenswrapper[4888]: I1124 01:09:54.536939 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4"} Nov 24 01:09:54 crc kubenswrapper[4888]: I1124 01:09:54.537650 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0"} Nov 24 01:09:54 crc kubenswrapper[4888]: I1124 01:09:54.537678 4888 scope.go:117] "RemoveContainer" containerID="e9c5b3230ab086a2f06360b4f0fb4d8ce719120b1196de4cd34f699db8d2d1a1" Nov 24 01:09:57 crc kubenswrapper[4888]: I1124 01:09:57.589025 4888 generic.go:334] "Generic (PLEG): container finished" podID="fdba4390-e81c-410b-9718-3d87e5ca4d82" containerID="ba0fd62114ef12169c9e8244b3c6d625df5019f48d7cffb5aab7cd8629eed080" exitCode=0 Nov 24 01:09:57 crc kubenswrapper[4888]: I1124 01:09:57.589144 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" event={"ID":"fdba4390-e81c-410b-9718-3d87e5ca4d82","Type":"ContainerDied","Data":"ba0fd62114ef12169c9e8244b3c6d625df5019f48d7cffb5aab7cd8629eed080"} Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.089779 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.223798 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.223922 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.224113 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.224171 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.224325 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.224428 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zslvl\" (UniqueName: \"kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.225201 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0\") pod \"fdba4390-e81c-410b-9718-3d87e5ca4d82\" (UID: \"fdba4390-e81c-410b-9718-3d87e5ca4d82\") " Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.232669 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl" (OuterVolumeSpecName: "kube-api-access-zslvl") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "kube-api-access-zslvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.233491 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.268360 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.282227 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.285090 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.286106 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory" (OuterVolumeSpecName: "inventory") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.288507 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "fdba4390-e81c-410b-9718-3d87e5ca4d82" (UID: "fdba4390-e81c-410b-9718-3d87e5ca4d82"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329617 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zslvl\" (UniqueName: \"kubernetes.io/projected/fdba4390-e81c-410b-9718-3d87e5ca4d82-kube-api-access-zslvl\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329647 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329660 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329674 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329686 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329699 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.329711 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/fdba4390-e81c-410b-9718-3d87e5ca4d82-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.629117 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" event={"ID":"fdba4390-e81c-410b-9718-3d87e5ca4d82","Type":"ContainerDied","Data":"1d9dc355926e87e804a695da62eb6262077008a8e036b025803bf41c728d8215"} Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.629181 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d9dc355926e87e804a695da62eb6262077008a8e036b025803bf41c728d8215" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.629216 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.808858 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h"] Nov 24 01:09:59 crc kubenswrapper[4888]: E1124 01:09:59.809253 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdba4390-e81c-410b-9718-3d87e5ca4d82" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809266 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdba4390-e81c-410b-9718-3d87e5ca4d82" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:09:59 crc kubenswrapper[4888]: E1124 01:09:59.809288 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="extract-content" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809295 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="extract-content" Nov 24 01:09:59 crc kubenswrapper[4888]: E1124 01:09:59.809319 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="extract-utilities" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809326 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="extract-utilities" Nov 24 01:09:59 crc kubenswrapper[4888]: E1124 01:09:59.809342 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="registry-server" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809347 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="registry-server" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809526 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdba4390-e81c-410b-9718-3d87e5ca4d82" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.809549 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa86846-b07e-45d9-b7db-eb4a088403b0" containerName="registry-server" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.810221 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.812851 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.813049 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.812963 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.813362 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.815062 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.837567 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h"] Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.942102 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.942175 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.942272 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.942400 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxgpj\" (UniqueName: \"kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:09:59 crc kubenswrapper[4888]: I1124 01:09:59.942597 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.044162 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.044270 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxgpj\" (UniqueName: \"kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.044369 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.044468 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.044498 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.049169 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.049178 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.049197 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.051442 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.061440 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxgpj\" (UniqueName: \"kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kbh4h\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:00 crc kubenswrapper[4888]: I1124 01:10:00.130458 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:01 crc kubenswrapper[4888]: I1124 01:10:01.082440 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h"] Nov 24 01:10:01 crc kubenswrapper[4888]: I1124 01:10:01.661368 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" event={"ID":"70c80f7c-35c2-418d-94eb-25927a0b6554","Type":"ContainerStarted","Data":"374442fad66b8715f393b12e7b6eadb395b18268388a9b8509eea3f18c4eaa35"} Nov 24 01:10:02 crc kubenswrapper[4888]: I1124 01:10:02.676919 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" event={"ID":"70c80f7c-35c2-418d-94eb-25927a0b6554","Type":"ContainerStarted","Data":"76eec6724dd183f13df63e6fb8ad6f6bafe519ab5dc6024a8228725ff9d33731"} Nov 24 01:10:02 crc kubenswrapper[4888]: I1124 01:10:02.721007 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" podStartSLOduration=3.12956396 podStartE2EDuration="3.720977247s" podCreationTimestamp="2025-11-24 01:09:59 +0000 UTC" firstStartedPulling="2025-11-24 01:10:01.09301128 +0000 UTC m=+2703.675695344" lastFinishedPulling="2025-11-24 01:10:01.684424547 +0000 UTC m=+2704.267108631" observedRunningTime="2025-11-24 01:10:02.706246922 +0000 UTC m=+2705.288930976" watchObservedRunningTime="2025-11-24 01:10:02.720977247 +0000 UTC m=+2705.303661301" Nov 24 01:10:22 crc kubenswrapper[4888]: I1124 01:10:22.896991 4888 generic.go:334] "Generic (PLEG): container finished" podID="70c80f7c-35c2-418d-94eb-25927a0b6554" containerID="76eec6724dd183f13df63e6fb8ad6f6bafe519ab5dc6024a8228725ff9d33731" exitCode=0 Nov 24 01:10:22 crc kubenswrapper[4888]: I1124 01:10:22.897054 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" event={"ID":"70c80f7c-35c2-418d-94eb-25927a0b6554","Type":"ContainerDied","Data":"76eec6724dd183f13df63e6fb8ad6f6bafe519ab5dc6024a8228725ff9d33731"} Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.400970 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.516876 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory\") pod \"70c80f7c-35c2-418d-94eb-25927a0b6554\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.517224 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0\") pod \"70c80f7c-35c2-418d-94eb-25927a0b6554\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.517314 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxgpj\" (UniqueName: \"kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj\") pod \"70c80f7c-35c2-418d-94eb-25927a0b6554\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.517429 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key\") pod \"70c80f7c-35c2-418d-94eb-25927a0b6554\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.517599 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1\") pod \"70c80f7c-35c2-418d-94eb-25927a0b6554\" (UID: \"70c80f7c-35c2-418d-94eb-25927a0b6554\") " Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.525307 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj" (OuterVolumeSpecName: "kube-api-access-bxgpj") pod "70c80f7c-35c2-418d-94eb-25927a0b6554" (UID: "70c80f7c-35c2-418d-94eb-25927a0b6554"). InnerVolumeSpecName "kube-api-access-bxgpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.557079 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70c80f7c-35c2-418d-94eb-25927a0b6554" (UID: "70c80f7c-35c2-418d-94eb-25927a0b6554"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.561523 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory" (OuterVolumeSpecName: "inventory") pod "70c80f7c-35c2-418d-94eb-25927a0b6554" (UID: "70c80f7c-35c2-418d-94eb-25927a0b6554"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.564688 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "70c80f7c-35c2-418d-94eb-25927a0b6554" (UID: "70c80f7c-35c2-418d-94eb-25927a0b6554"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.575442 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "70c80f7c-35c2-418d-94eb-25927a0b6554" (UID: "70c80f7c-35c2-418d-94eb-25927a0b6554"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.620616 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.620653 4888 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.620667 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxgpj\" (UniqueName: \"kubernetes.io/projected/70c80f7c-35c2-418d-94eb-25927a0b6554-kube-api-access-bxgpj\") on node \"crc\" DevicePath \"\"" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.620680 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.620692 4888 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/70c80f7c-35c2-418d-94eb-25927a0b6554-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.920361 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" event={"ID":"70c80f7c-35c2-418d-94eb-25927a0b6554","Type":"ContainerDied","Data":"374442fad66b8715f393b12e7b6eadb395b18268388a9b8509eea3f18c4eaa35"} Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.920404 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="374442fad66b8715f393b12e7b6eadb395b18268388a9b8509eea3f18c4eaa35" Nov 24 01:10:24 crc kubenswrapper[4888]: I1124 01:10:24.920495 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.119461 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:18 crc kubenswrapper[4888]: E1124 01:11:18.121120 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c80f7c-35c2-418d-94eb-25927a0b6554" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.121159 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c80f7c-35c2-418d-94eb-25927a0b6554" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.121580 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c80f7c-35c2-418d-94eb-25927a0b6554" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.124804 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.130060 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.271756 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.272243 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.272464 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjk9z\" (UniqueName: \"kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.375107 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.375582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjk9z\" (UniqueName: \"kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.375627 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.375802 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.376551 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.404186 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjk9z\" (UniqueName: \"kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z\") pod \"redhat-operators-wk924\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.459902 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:18 crc kubenswrapper[4888]: I1124 01:11:18.930020 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:19 crc kubenswrapper[4888]: I1124 01:11:19.585304 4888 generic.go:334] "Generic (PLEG): container finished" podID="1f644255-eb41-4226-a1bd-e550b28c8823" containerID="9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82" exitCode=0 Nov 24 01:11:19 crc kubenswrapper[4888]: I1124 01:11:19.585374 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerDied","Data":"9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82"} Nov 24 01:11:19 crc kubenswrapper[4888]: I1124 01:11:19.585669 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerStarted","Data":"a32a2abd0312cfd0139e2c8c0a5cd0450ac9bfd621e6849d11d5fec4d3a41dde"} Nov 24 01:11:19 crc kubenswrapper[4888]: I1124 01:11:19.588680 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:11:21 crc kubenswrapper[4888]: I1124 01:11:21.611114 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerStarted","Data":"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa"} Nov 24 01:11:23 crc kubenswrapper[4888]: I1124 01:11:23.634953 4888 generic.go:334] "Generic (PLEG): container finished" podID="1f644255-eb41-4226-a1bd-e550b28c8823" containerID="a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa" exitCode=0 Nov 24 01:11:23 crc kubenswrapper[4888]: I1124 01:11:23.635071 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerDied","Data":"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa"} Nov 24 01:11:24 crc kubenswrapper[4888]: I1124 01:11:24.647947 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerStarted","Data":"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e"} Nov 24 01:11:24 crc kubenswrapper[4888]: I1124 01:11:24.668088 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wk924" podStartSLOduration=2.207480857 podStartE2EDuration="6.668062658s" podCreationTimestamp="2025-11-24 01:11:18 +0000 UTC" firstStartedPulling="2025-11-24 01:11:19.588432943 +0000 UTC m=+2782.171116987" lastFinishedPulling="2025-11-24 01:11:24.049014714 +0000 UTC m=+2786.631698788" observedRunningTime="2025-11-24 01:11:24.663015756 +0000 UTC m=+2787.245699840" watchObservedRunningTime="2025-11-24 01:11:24.668062658 +0000 UTC m=+2787.250746712" Nov 24 01:11:28 crc kubenswrapper[4888]: I1124 01:11:28.460724 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:28 crc kubenswrapper[4888]: I1124 01:11:28.461553 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:29 crc kubenswrapper[4888]: I1124 01:11:29.521648 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wk924" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="registry-server" probeResult="failure" output=< Nov 24 01:11:29 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 01:11:29 crc kubenswrapper[4888]: > Nov 24 01:11:38 crc kubenswrapper[4888]: I1124 01:11:38.545660 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:38 crc kubenswrapper[4888]: I1124 01:11:38.606608 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:38 crc kubenswrapper[4888]: I1124 01:11:38.796062 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:39 crc kubenswrapper[4888]: I1124 01:11:39.826892 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wk924" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="registry-server" containerID="cri-o://0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e" gracePeriod=2 Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.354658 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.466253 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content\") pod \"1f644255-eb41-4226-a1bd-e550b28c8823\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.466519 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities\") pod \"1f644255-eb41-4226-a1bd-e550b28c8823\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.466687 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjk9z\" (UniqueName: \"kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z\") pod \"1f644255-eb41-4226-a1bd-e550b28c8823\" (UID: \"1f644255-eb41-4226-a1bd-e550b28c8823\") " Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.467382 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities" (OuterVolumeSpecName: "utilities") pod "1f644255-eb41-4226-a1bd-e550b28c8823" (UID: "1f644255-eb41-4226-a1bd-e550b28c8823"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.468619 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.475583 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z" (OuterVolumeSpecName: "kube-api-access-sjk9z") pod "1f644255-eb41-4226-a1bd-e550b28c8823" (UID: "1f644255-eb41-4226-a1bd-e550b28c8823"). InnerVolumeSpecName "kube-api-access-sjk9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.571037 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjk9z\" (UniqueName: \"kubernetes.io/projected/1f644255-eb41-4226-a1bd-e550b28c8823-kube-api-access-sjk9z\") on node \"crc\" DevicePath \"\"" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.577291 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f644255-eb41-4226-a1bd-e550b28c8823" (UID: "1f644255-eb41-4226-a1bd-e550b28c8823"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.673157 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f644255-eb41-4226-a1bd-e550b28c8823-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.841632 4888 generic.go:334] "Generic (PLEG): container finished" podID="1f644255-eb41-4226-a1bd-e550b28c8823" containerID="0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e" exitCode=0 Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.841698 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerDied","Data":"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e"} Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.841741 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk924" event={"ID":"1f644255-eb41-4226-a1bd-e550b28c8823","Type":"ContainerDied","Data":"a32a2abd0312cfd0139e2c8c0a5cd0450ac9bfd621e6849d11d5fec4d3a41dde"} Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.841776 4888 scope.go:117] "RemoveContainer" containerID="0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.841777 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk924" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.869610 4888 scope.go:117] "RemoveContainer" containerID="a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.898095 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.908322 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wk924"] Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.927344 4888 scope.go:117] "RemoveContainer" containerID="9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.966086 4888 scope.go:117] "RemoveContainer" containerID="0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e" Nov 24 01:11:40 crc kubenswrapper[4888]: E1124 01:11:40.966513 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e\": container with ID starting with 0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e not found: ID does not exist" containerID="0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.966542 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e"} err="failed to get container status \"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e\": rpc error: code = NotFound desc = could not find container \"0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e\": container with ID starting with 0019e1366af3dd509395e56ef497be5cf953977d22b3957f7921a575b630c91e not found: ID does not exist" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.966565 4888 scope.go:117] "RemoveContainer" containerID="a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa" Nov 24 01:11:40 crc kubenswrapper[4888]: E1124 01:11:40.966839 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa\": container with ID starting with a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa not found: ID does not exist" containerID="a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.966893 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa"} err="failed to get container status \"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa\": rpc error: code = NotFound desc = could not find container \"a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa\": container with ID starting with a466fb8e26f015ee346880ca2c412f30803c7390f08efa152469144b01c354aa not found: ID does not exist" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.966931 4888 scope.go:117] "RemoveContainer" containerID="9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82" Nov 24 01:11:40 crc kubenswrapper[4888]: E1124 01:11:40.967443 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82\": container with ID starting with 9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82 not found: ID does not exist" containerID="9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82" Nov 24 01:11:40 crc kubenswrapper[4888]: I1124 01:11:40.967522 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82"} err="failed to get container status \"9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82\": rpc error: code = NotFound desc = could not find container \"9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82\": container with ID starting with 9ed248bb192b161f98ec02d4728dbc7516a62577245193a1c20115dc7f21ce82 not found: ID does not exist" Nov 24 01:11:42 crc kubenswrapper[4888]: I1124 01:11:42.270192 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" path="/var/lib/kubelet/pods/1f644255-eb41-4226-a1bd-e550b28c8823/volumes" Nov 24 01:11:53 crc kubenswrapper[4888]: I1124 01:11:53.591303 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:11:53 crc kubenswrapper[4888]: I1124 01:11:53.592204 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:12:23 crc kubenswrapper[4888]: I1124 01:12:23.591275 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:12:23 crc kubenswrapper[4888]: I1124 01:12:23.592005 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.591343 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.591777 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.591830 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.592560 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.592611 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" gracePeriod=600 Nov 24 01:12:53 crc kubenswrapper[4888]: E1124 01:12:53.732225 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.750753 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" exitCode=0 Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.750841 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0"} Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.750904 4888 scope.go:117] "RemoveContainer" containerID="f2fc8e461de112a3a1caa7171c1d61ef0a69185ae9de48397e6e89a468bfc5d4" Nov 24 01:12:53 crc kubenswrapper[4888]: I1124 01:12:53.751760 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:12:53 crc kubenswrapper[4888]: E1124 01:12:53.752096 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:13:04 crc kubenswrapper[4888]: I1124 01:13:04.246050 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:13:04 crc kubenswrapper[4888]: E1124 01:13:04.246838 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:13:18 crc kubenswrapper[4888]: I1124 01:13:18.258419 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:13:18 crc kubenswrapper[4888]: E1124 01:13:18.259678 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:13:31 crc kubenswrapper[4888]: I1124 01:13:31.246162 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:13:31 crc kubenswrapper[4888]: E1124 01:13:31.247657 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:13:45 crc kubenswrapper[4888]: I1124 01:13:45.245936 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:13:45 crc kubenswrapper[4888]: E1124 01:13:45.247263 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:14:00 crc kubenswrapper[4888]: I1124 01:14:00.246199 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:14:00 crc kubenswrapper[4888]: E1124 01:14:00.247252 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:14:12 crc kubenswrapper[4888]: I1124 01:14:12.246144 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:14:12 crc kubenswrapper[4888]: E1124 01:14:12.250619 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:14:23 crc kubenswrapper[4888]: I1124 01:14:23.246844 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:14:23 crc kubenswrapper[4888]: E1124 01:14:23.247556 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:14:36 crc kubenswrapper[4888]: I1124 01:14:36.245887 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:14:36 crc kubenswrapper[4888]: E1124 01:14:36.246789 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:14:48 crc kubenswrapper[4888]: I1124 01:14:48.253666 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:14:48 crc kubenswrapper[4888]: E1124 01:14:48.254986 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.182034 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls"] Nov 24 01:15:00 crc kubenswrapper[4888]: E1124 01:15:00.183465 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="extract-utilities" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.183481 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="extract-utilities" Nov 24 01:15:00 crc kubenswrapper[4888]: E1124 01:15:00.183504 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="registry-server" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.183510 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="registry-server" Nov 24 01:15:00 crc kubenswrapper[4888]: E1124 01:15:00.183536 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="extract-content" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.183543 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="extract-content" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.183861 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f644255-eb41-4226-a1bd-e550b28c8823" containerName="registry-server" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.184692 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.186953 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.188032 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.192284 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls"] Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.245769 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:15:00 crc kubenswrapper[4888]: E1124 01:15:00.246084 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.275123 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.275342 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj44l\" (UniqueName: \"kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.275457 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.377660 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj44l\" (UniqueName: \"kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.377770 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.377844 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.378882 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.385495 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.394974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj44l\" (UniqueName: \"kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l\") pod \"collect-profiles-29399115-z9tls\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:00 crc kubenswrapper[4888]: I1124 01:15:00.510110 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:01 crc kubenswrapper[4888]: I1124 01:15:01.003862 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls"] Nov 24 01:15:01 crc kubenswrapper[4888]: I1124 01:15:01.450326 4888 generic.go:334] "Generic (PLEG): container finished" podID="61ae0db9-fa05-4603-b9f8-7ff49fa36022" containerID="c6a5a5bb86a990bee92b608a77f9a4c69b77793503ad55fad09a4aabbb065f55" exitCode=0 Nov 24 01:15:01 crc kubenswrapper[4888]: I1124 01:15:01.450378 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" event={"ID":"61ae0db9-fa05-4603-b9f8-7ff49fa36022","Type":"ContainerDied","Data":"c6a5a5bb86a990bee92b608a77f9a4c69b77793503ad55fad09a4aabbb065f55"} Nov 24 01:15:01 crc kubenswrapper[4888]: I1124 01:15:01.450748 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" event={"ID":"61ae0db9-fa05-4603-b9f8-7ff49fa36022","Type":"ContainerStarted","Data":"70d804bd7255bfd25f5e997ff9f0b804f9656e44c75297851d11288d106ec7dc"} Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.835155 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.936092 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj44l\" (UniqueName: \"kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l\") pod \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.936243 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume\") pod \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.936282 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume\") pod \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\" (UID: \"61ae0db9-fa05-4603-b9f8-7ff49fa36022\") " Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.936983 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume" (OuterVolumeSpecName: "config-volume") pod "61ae0db9-fa05-4603-b9f8-7ff49fa36022" (UID: "61ae0db9-fa05-4603-b9f8-7ff49fa36022"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.941770 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "61ae0db9-fa05-4603-b9f8-7ff49fa36022" (UID: "61ae0db9-fa05-4603-b9f8-7ff49fa36022"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:15:02 crc kubenswrapper[4888]: I1124 01:15:02.943164 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l" (OuterVolumeSpecName: "kube-api-access-jj44l") pod "61ae0db9-fa05-4603-b9f8-7ff49fa36022" (UID: "61ae0db9-fa05-4603-b9f8-7ff49fa36022"). InnerVolumeSpecName "kube-api-access-jj44l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.037608 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61ae0db9-fa05-4603-b9f8-7ff49fa36022-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.037644 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj44l\" (UniqueName: \"kubernetes.io/projected/61ae0db9-fa05-4603-b9f8-7ff49fa36022-kube-api-access-jj44l\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.037653 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61ae0db9-fa05-4603-b9f8-7ff49fa36022-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.490750 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" event={"ID":"61ae0db9-fa05-4603-b9f8-7ff49fa36022","Type":"ContainerDied","Data":"70d804bd7255bfd25f5e997ff9f0b804f9656e44c75297851d11288d106ec7dc"} Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.491104 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70d804bd7255bfd25f5e997ff9f0b804f9656e44c75297851d11288d106ec7dc" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.490884 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls" Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.917364 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q"] Nov 24 01:15:03 crc kubenswrapper[4888]: I1124 01:15:03.927983 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399070-pvx4q"] Nov 24 01:15:04 crc kubenswrapper[4888]: I1124 01:15:04.260941 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0282dd1d-4a52-4986-a79c-01d365dab893" path="/var/lib/kubelet/pods/0282dd1d-4a52-4986-a79c-01d365dab893/volumes" Nov 24 01:15:12 crc kubenswrapper[4888]: I1124 01:15:12.652013 4888 scope.go:117] "RemoveContainer" containerID="7a4a5bec4038ef1dfcec7a467b296167de8a934cb72f2cc9b9f80c6829a521f6" Nov 24 01:15:14 crc kubenswrapper[4888]: I1124 01:15:14.246066 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:15:14 crc kubenswrapper[4888]: E1124 01:15:14.247164 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.350227 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.359396 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.406298 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.421017 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.429219 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.438219 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v8tnj"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.447829 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zjdzq"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.456553 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-xrbrm"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.464643 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tfbqn"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.473246 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.482369 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s62w2"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.490897 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.499646 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-74w7m"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.509396 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kbh4h"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.517448 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.527372 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.536334 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.544761 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sl6ml"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.552303 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbj6r"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.559726 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qdqdf"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.566665 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.573288 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mwsbj"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.580403 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.587396 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.594714 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hsxgp"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.601271 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.607544 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.613655 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qdqdf"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.619785 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jdndl"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.626484 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gjfzs"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.633949 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-d4lsr"] Nov 24 01:15:26 crc kubenswrapper[4888]: I1124 01:15:26.640441 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8jk4"] Nov 24 01:15:27 crc kubenswrapper[4888]: I1124 01:15:27.245938 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:15:27 crc kubenswrapper[4888]: E1124 01:15:27.246861 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.261836 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089fe5dd-64b5-4982-aa29-4c8025a04581" path="/var/lib/kubelet/pods/089fe5dd-64b5-4982-aa29-4c8025a04581/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.262995 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8f2d24-17cd-4aed-b5f3-b71ef305b58b" path="/var/lib/kubelet/pods/1d8f2d24-17cd-4aed-b5f3-b71ef305b58b/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.264199 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="371f2fbd-a533-4ac3-85e8-68b5c3dccc25" path="/var/lib/kubelet/pods/371f2fbd-a533-4ac3-85e8-68b5c3dccc25/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.265284 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f89ceb2-718c-4ae5-bd9c-7b518d550cd3" path="/var/lib/kubelet/pods/3f89ceb2-718c-4ae5-bd9c-7b518d550cd3/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.267596 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a52b97b-92bd-42ab-8b5b-80eafbdce0fa" path="/var/lib/kubelet/pods/4a52b97b-92bd-42ab-8b5b-80eafbdce0fa/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.268375 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9" path="/var/lib/kubelet/pods/4a7bf02f-e770-43a9-8c1d-63d83a3d7bb9/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.269070 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c6eda4-874e-479b-b8cc-2f9212a197f8" path="/var/lib/kubelet/pods/52c6eda4-874e-479b-b8cc-2f9212a197f8/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.271208 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64895ae8-ff60-4598-b404-cdd6c14c0c02" path="/var/lib/kubelet/pods/64895ae8-ff60-4598-b404-cdd6c14c0c02/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.271934 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67f5d5b8-4bdd-4601-978f-30a1dff053c0" path="/var/lib/kubelet/pods/67f5d5b8-4bdd-4601-978f-30a1dff053c0/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.272613 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c80f7c-35c2-418d-94eb-25927a0b6554" path="/var/lib/kubelet/pods/70c80f7c-35c2-418d-94eb-25927a0b6554/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.273924 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94505667-e4ca-4420-bb6b-bff063bab9ec" path="/var/lib/kubelet/pods/94505667-e4ca-4420-bb6b-bff063bab9ec/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.274567 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a90f3597-c206-488b-a833-49c9dd906ac9" path="/var/lib/kubelet/pods/a90f3597-c206-488b-a833-49c9dd906ac9/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.275308 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca5b9705-eedc-46e0-9f41-4b070305f441" path="/var/lib/kubelet/pods/ca5b9705-eedc-46e0-9f41-4b070305f441/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.275943 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f58fbdbc-1266-410e-b2fb-2e1aec45104e" path="/var/lib/kubelet/pods/f58fbdbc-1266-410e-b2fb-2e1aec45104e/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.277072 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e9a324-c15a-4b2e-8d60-4651313aca06" path="/var/lib/kubelet/pods/f5e9a324-c15a-4b2e-8d60-4651313aca06/volumes" Nov 24 01:15:28 crc kubenswrapper[4888]: I1124 01:15:28.277623 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdba4390-e81c-410b-9718-3d87e5ca4d82" path="/var/lib/kubelet/pods/fdba4390-e81c-410b-9718-3d87e5ca4d82/volumes" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.482362 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt"] Nov 24 01:15:31 crc kubenswrapper[4888]: E1124 01:15:31.483786 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61ae0db9-fa05-4603-b9f8-7ff49fa36022" containerName="collect-profiles" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.483834 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="61ae0db9-fa05-4603-b9f8-7ff49fa36022" containerName="collect-profiles" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.484192 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="61ae0db9-fa05-4603-b9f8-7ff49fa36022" containerName="collect-profiles" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.485374 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.488934 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.489880 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.489930 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.490123 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.490315 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.493637 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt"] Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.570048 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.570308 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.570416 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.570528 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lzwp\" (UniqueName: \"kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.570613 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.672442 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.672518 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.672565 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.672652 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lzwp\" (UniqueName: \"kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.672708 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.679600 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.680238 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.680249 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.680584 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.691457 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lzwp\" (UniqueName: \"kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:31 crc kubenswrapper[4888]: I1124 01:15:31.811786 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:32 crc kubenswrapper[4888]: I1124 01:15:32.497729 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt"] Nov 24 01:15:32 crc kubenswrapper[4888]: I1124 01:15:32.796319 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" event={"ID":"30e410dc-697f-4714-9e4b-a2f8a71796c4","Type":"ContainerStarted","Data":"a16cdc5bff536f1fa2394eacb3deab2aa3f3106e3faa503ed8926b208cfe2d60"} Nov 24 01:15:33 crc kubenswrapper[4888]: I1124 01:15:33.809497 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" event={"ID":"30e410dc-697f-4714-9e4b-a2f8a71796c4","Type":"ContainerStarted","Data":"f826fea96a7e2c8d164cc5e887af23b672dfa2bb142eb0b1146501392182c371"} Nov 24 01:15:33 crc kubenswrapper[4888]: I1124 01:15:33.837112 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" podStartSLOduration=2.299629027 podStartE2EDuration="2.837085236s" podCreationTimestamp="2025-11-24 01:15:31 +0000 UTC" firstStartedPulling="2025-11-24 01:15:32.515771867 +0000 UTC m=+3035.098455911" lastFinishedPulling="2025-11-24 01:15:33.053228076 +0000 UTC m=+3035.635912120" observedRunningTime="2025-11-24 01:15:33.83261594 +0000 UTC m=+3036.415300024" watchObservedRunningTime="2025-11-24 01:15:33.837085236 +0000 UTC m=+3036.419769310" Nov 24 01:15:42 crc kubenswrapper[4888]: I1124 01:15:42.246043 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:15:42 crc kubenswrapper[4888]: E1124 01:15:42.250789 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:15:45 crc kubenswrapper[4888]: I1124 01:15:45.960048 4888 generic.go:334] "Generic (PLEG): container finished" podID="30e410dc-697f-4714-9e4b-a2f8a71796c4" containerID="f826fea96a7e2c8d164cc5e887af23b672dfa2bb142eb0b1146501392182c371" exitCode=0 Nov 24 01:15:45 crc kubenswrapper[4888]: I1124 01:15:45.960155 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" event={"ID":"30e410dc-697f-4714-9e4b-a2f8a71796c4","Type":"ContainerDied","Data":"f826fea96a7e2c8d164cc5e887af23b672dfa2bb142eb0b1146501392182c371"} Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.459301 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.547782 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle\") pod \"30e410dc-697f-4714-9e4b-a2f8a71796c4\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.547884 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key\") pod \"30e410dc-697f-4714-9e4b-a2f8a71796c4\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.548001 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory\") pod \"30e410dc-697f-4714-9e4b-a2f8a71796c4\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.548115 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lzwp\" (UniqueName: \"kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp\") pod \"30e410dc-697f-4714-9e4b-a2f8a71796c4\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.548186 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph\") pod \"30e410dc-697f-4714-9e4b-a2f8a71796c4\" (UID: \"30e410dc-697f-4714-9e4b-a2f8a71796c4\") " Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.553729 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph" (OuterVolumeSpecName: "ceph") pod "30e410dc-697f-4714-9e4b-a2f8a71796c4" (UID: "30e410dc-697f-4714-9e4b-a2f8a71796c4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.556110 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp" (OuterVolumeSpecName: "kube-api-access-2lzwp") pod "30e410dc-697f-4714-9e4b-a2f8a71796c4" (UID: "30e410dc-697f-4714-9e4b-a2f8a71796c4"). InnerVolumeSpecName "kube-api-access-2lzwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.556407 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "30e410dc-697f-4714-9e4b-a2f8a71796c4" (UID: "30e410dc-697f-4714-9e4b-a2f8a71796c4"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.586161 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "30e410dc-697f-4714-9e4b-a2f8a71796c4" (UID: "30e410dc-697f-4714-9e4b-a2f8a71796c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.605052 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory" (OuterVolumeSpecName: "inventory") pod "30e410dc-697f-4714-9e4b-a2f8a71796c4" (UID: "30e410dc-697f-4714-9e4b-a2f8a71796c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.650784 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lzwp\" (UniqueName: \"kubernetes.io/projected/30e410dc-697f-4714-9e4b-a2f8a71796c4-kube-api-access-2lzwp\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.650816 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.650840 4888 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.650856 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.650864 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30e410dc-697f-4714-9e4b-a2f8a71796c4-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.980582 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" event={"ID":"30e410dc-697f-4714-9e4b-a2f8a71796c4","Type":"ContainerDied","Data":"a16cdc5bff536f1fa2394eacb3deab2aa3f3106e3faa503ed8926b208cfe2d60"} Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.980624 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a16cdc5bff536f1fa2394eacb3deab2aa3f3106e3faa503ed8926b208cfe2d60" Nov 24 01:15:47 crc kubenswrapper[4888]: I1124 01:15:47.980645 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.190969 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt"] Nov 24 01:15:48 crc kubenswrapper[4888]: E1124 01:15:48.191444 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e410dc-697f-4714-9e4b-a2f8a71796c4" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.191468 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e410dc-697f-4714-9e4b-a2f8a71796c4" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.191731 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e410dc-697f-4714-9e4b-a2f8a71796c4" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.192533 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.195044 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.195170 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.195562 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.195864 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.196144 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.211378 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt"] Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.260761 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.260841 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.260964 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.260993 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.261063 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv4tp\" (UniqueName: \"kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.363583 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.363701 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.363863 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv4tp\" (UniqueName: \"kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.364046 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.364104 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.369161 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.369437 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.369444 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.371027 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.380778 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv4tp\" (UniqueName: \"kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:48 crc kubenswrapper[4888]: I1124 01:15:48.509885 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:15:49 crc kubenswrapper[4888]: I1124 01:15:49.064398 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt"] Nov 24 01:15:49 crc kubenswrapper[4888]: W1124 01:15:49.066489 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebbc9282_a85b_4484_b130_5b0c91030009.slice/crio-d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3 WatchSource:0}: Error finding container d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3: Status 404 returned error can't find the container with id d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3 Nov 24 01:15:50 crc kubenswrapper[4888]: I1124 01:15:50.008095 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" event={"ID":"ebbc9282-a85b-4484-b130-5b0c91030009","Type":"ContainerStarted","Data":"389c0e903db9b9cff159f36e8c48c9bfaa6a78d97e49ee6fab54352f30683e2c"} Nov 24 01:15:50 crc kubenswrapper[4888]: I1124 01:15:50.008442 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" event={"ID":"ebbc9282-a85b-4484-b130-5b0c91030009","Type":"ContainerStarted","Data":"d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3"} Nov 24 01:15:50 crc kubenswrapper[4888]: I1124 01:15:50.029349 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" podStartSLOduration=1.618843311 podStartE2EDuration="2.029323572s" podCreationTimestamp="2025-11-24 01:15:48 +0000 UTC" firstStartedPulling="2025-11-24 01:15:49.068914547 +0000 UTC m=+3051.651598601" lastFinishedPulling="2025-11-24 01:15:49.479394778 +0000 UTC m=+3052.062078862" observedRunningTime="2025-11-24 01:15:50.027374057 +0000 UTC m=+3052.610058121" watchObservedRunningTime="2025-11-24 01:15:50.029323572 +0000 UTC m=+3052.612007636" Nov 24 01:15:53 crc kubenswrapper[4888]: I1124 01:15:53.245559 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:15:53 crc kubenswrapper[4888]: E1124 01:15:53.246218 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:16:06 crc kubenswrapper[4888]: I1124 01:16:06.245547 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:16:06 crc kubenswrapper[4888]: E1124 01:16:06.246393 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:16:12 crc kubenswrapper[4888]: I1124 01:16:12.847883 4888 scope.go:117] "RemoveContainer" containerID="de61341608f4f48f79daf8dc32d4f5cfd4e0e323f804295b47ed98a67e17f959" Nov 24 01:16:12 crc kubenswrapper[4888]: I1124 01:16:12.924119 4888 scope.go:117] "RemoveContainer" containerID="0e20725fc483a9c7ea2bde0a0c5d5c4941ea60ed74c5ee28194d1d6dc528995d" Nov 24 01:16:12 crc kubenswrapper[4888]: I1124 01:16:12.966584 4888 scope.go:117] "RemoveContainer" containerID="ba0fd62114ef12169c9e8244b3c6d625df5019f48d7cffb5aab7cd8629eed080" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.074611 4888 scope.go:117] "RemoveContainer" containerID="70fcba0c6aa150fb3696f379be7fdab98b6cd1006b50926a9b2b1339ca6610c8" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.121724 4888 scope.go:117] "RemoveContainer" containerID="91052a71ca06d6c79bd06daeeede9f8264d95969de631793683b516d1532de21" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.212873 4888 scope.go:117] "RemoveContainer" containerID="0ece25aa8b97abdb3d29e524bf4bd33c7d0ca106dfc9f77fcdc10cbcfed948aa" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.271260 4888 scope.go:117] "RemoveContainer" containerID="89489d7a4ece5c2081b8ce17e9e3c7a1a376579875933344d8b27632e73194c7" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.309180 4888 scope.go:117] "RemoveContainer" containerID="76eec6724dd183f13df63e6fb8ad6f6bafe519ab5dc6024a8228725ff9d33731" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.353171 4888 scope.go:117] "RemoveContainer" containerID="92af64c26868dc23429e73d4be9e712469f06ed7274b2a50f9b2c5394427ad0e" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.429252 4888 scope.go:117] "RemoveContainer" containerID="0c936c71c2b08c6db064838ebc13a00e9a15210df527ffcc121fa5dc71395dc1" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.469284 4888 scope.go:117] "RemoveContainer" containerID="bc673cba94d2aa570b541e34c3178d7e6bcb000739417b668b26ec567bb817e0" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.548464 4888 scope.go:117] "RemoveContainer" containerID="840c9406f811f5d10bf819048ed255dcc7b6891c3b8028c5ef40822f7c5d560f" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.615326 4888 scope.go:117] "RemoveContainer" containerID="37173908abef59b4695d933068c59d24092cb8036e13dc4b8a72d6806c7daef6" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.707617 4888 scope.go:117] "RemoveContainer" containerID="1bb0be91508bf755e4688fb9709896762542e4c6ef528b6e98303296dc504159" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.749884 4888 scope.go:117] "RemoveContainer" containerID="4012e3bc40a0281cd50745adc810301ba39a7869517d93044d2bffaa830d9bcc" Nov 24 01:16:13 crc kubenswrapper[4888]: I1124 01:16:13.782949 4888 scope.go:117] "RemoveContainer" containerID="1c4855ea7bb8c5e2544a960adaa722e37260afe655c073d0f2be83032156fa78" Nov 24 01:16:17 crc kubenswrapper[4888]: I1124 01:16:17.244844 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:16:17 crc kubenswrapper[4888]: E1124 01:16:17.245476 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:16:30 crc kubenswrapper[4888]: I1124 01:16:30.245319 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:16:30 crc kubenswrapper[4888]: E1124 01:16:30.246144 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.734163 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.738906 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.753558 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.834470 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snwf4\" (UniqueName: \"kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.834608 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.834785 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.935785 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.935879 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snwf4\" (UniqueName: \"kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.935945 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.936302 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.936480 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:36 crc kubenswrapper[4888]: I1124 01:16:36.961545 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snwf4\" (UniqueName: \"kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4\") pod \"community-operators-h9dwq\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:37 crc kubenswrapper[4888]: I1124 01:16:37.075205 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:37 crc kubenswrapper[4888]: I1124 01:16:37.624686 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:38 crc kubenswrapper[4888]: I1124 01:16:38.599901 4888 generic.go:334] "Generic (PLEG): container finished" podID="1c310cbf-a058-4590-bd21-45eea8198d35" containerID="4c5bf4d62a53585e9ccb5f751c8c202752965c9809c3409595e0c33f78f07118" exitCode=0 Nov 24 01:16:38 crc kubenswrapper[4888]: I1124 01:16:38.599949 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerDied","Data":"4c5bf4d62a53585e9ccb5f751c8c202752965c9809c3409595e0c33f78f07118"} Nov 24 01:16:38 crc kubenswrapper[4888]: I1124 01:16:38.599979 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerStarted","Data":"102c0b68ad81d75d8795ade2a5d157a7ded2aa7655c06ebe1e37255efd8edfd6"} Nov 24 01:16:38 crc kubenswrapper[4888]: I1124 01:16:38.604222 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:16:39 crc kubenswrapper[4888]: I1124 01:16:39.611100 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerStarted","Data":"bac6aac4dacbd78b69913fa4fcf2433d905a370d9406f2b306700347e3c907cf"} Nov 24 01:16:40 crc kubenswrapper[4888]: I1124 01:16:40.632488 4888 generic.go:334] "Generic (PLEG): container finished" podID="1c310cbf-a058-4590-bd21-45eea8198d35" containerID="bac6aac4dacbd78b69913fa4fcf2433d905a370d9406f2b306700347e3c907cf" exitCode=0 Nov 24 01:16:40 crc kubenswrapper[4888]: I1124 01:16:40.632552 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerDied","Data":"bac6aac4dacbd78b69913fa4fcf2433d905a370d9406f2b306700347e3c907cf"} Nov 24 01:16:41 crc kubenswrapper[4888]: I1124 01:16:41.647102 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerStarted","Data":"a6bffa3456299327e90be492361c2c14c78a33ba026b7f15ee91575dc0a8b3de"} Nov 24 01:16:41 crc kubenswrapper[4888]: I1124 01:16:41.668821 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h9dwq" podStartSLOduration=3.262150881 podStartE2EDuration="5.66878778s" podCreationTimestamp="2025-11-24 01:16:36 +0000 UTC" firstStartedPulling="2025-11-24 01:16:38.603782936 +0000 UTC m=+3101.186466990" lastFinishedPulling="2025-11-24 01:16:41.010419845 +0000 UTC m=+3103.593103889" observedRunningTime="2025-11-24 01:16:41.663401348 +0000 UTC m=+3104.246085432" watchObservedRunningTime="2025-11-24 01:16:41.66878778 +0000 UTC m=+3104.251471834" Nov 24 01:16:44 crc kubenswrapper[4888]: I1124 01:16:44.246498 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:16:44 crc kubenswrapper[4888]: E1124 01:16:44.247491 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:16:47 crc kubenswrapper[4888]: I1124 01:16:47.076046 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:47 crc kubenswrapper[4888]: I1124 01:16:47.076461 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:47 crc kubenswrapper[4888]: I1124 01:16:47.163275 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:47 crc kubenswrapper[4888]: I1124 01:16:47.783355 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:47 crc kubenswrapper[4888]: I1124 01:16:47.841269 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:49 crc kubenswrapper[4888]: I1124 01:16:49.750727 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h9dwq" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="registry-server" containerID="cri-o://a6bffa3456299327e90be492361c2c14c78a33ba026b7f15ee91575dc0a8b3de" gracePeriod=2 Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.768171 4888 generic.go:334] "Generic (PLEG): container finished" podID="1c310cbf-a058-4590-bd21-45eea8198d35" containerID="a6bffa3456299327e90be492361c2c14c78a33ba026b7f15ee91575dc0a8b3de" exitCode=0 Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.768774 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerDied","Data":"a6bffa3456299327e90be492361c2c14c78a33ba026b7f15ee91575dc0a8b3de"} Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.769504 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9dwq" event={"ID":"1c310cbf-a058-4590-bd21-45eea8198d35","Type":"ContainerDied","Data":"102c0b68ad81d75d8795ade2a5d157a7ded2aa7655c06ebe1e37255efd8edfd6"} Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.769567 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="102c0b68ad81d75d8795ade2a5d157a7ded2aa7655c06ebe1e37255efd8edfd6" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.780733 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.875149 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snwf4\" (UniqueName: \"kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4\") pod \"1c310cbf-a058-4590-bd21-45eea8198d35\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.875380 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities\") pod \"1c310cbf-a058-4590-bd21-45eea8198d35\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.875712 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content\") pod \"1c310cbf-a058-4590-bd21-45eea8198d35\" (UID: \"1c310cbf-a058-4590-bd21-45eea8198d35\") " Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.876289 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities" (OuterVolumeSpecName: "utilities") pod "1c310cbf-a058-4590-bd21-45eea8198d35" (UID: "1c310cbf-a058-4590-bd21-45eea8198d35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.876880 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.881747 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4" (OuterVolumeSpecName: "kube-api-access-snwf4") pod "1c310cbf-a058-4590-bd21-45eea8198d35" (UID: "1c310cbf-a058-4590-bd21-45eea8198d35"). InnerVolumeSpecName "kube-api-access-snwf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.957935 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c310cbf-a058-4590-bd21-45eea8198d35" (UID: "1c310cbf-a058-4590-bd21-45eea8198d35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.981502 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snwf4\" (UniqueName: \"kubernetes.io/projected/1c310cbf-a058-4590-bd21-45eea8198d35-kube-api-access-snwf4\") on node \"crc\" DevicePath \"\"" Nov 24 01:16:50 crc kubenswrapper[4888]: I1124 01:16:50.981549 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c310cbf-a058-4590-bd21-45eea8198d35-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:16:51 crc kubenswrapper[4888]: I1124 01:16:51.784435 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9dwq" Nov 24 01:16:51 crc kubenswrapper[4888]: I1124 01:16:51.847779 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:51 crc kubenswrapper[4888]: I1124 01:16:51.863045 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h9dwq"] Nov 24 01:16:52 crc kubenswrapper[4888]: I1124 01:16:52.271052 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" path="/var/lib/kubelet/pods/1c310cbf-a058-4590-bd21-45eea8198d35/volumes" Nov 24 01:16:57 crc kubenswrapper[4888]: I1124 01:16:57.245653 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:16:57 crc kubenswrapper[4888]: E1124 01:16:57.247222 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:17:09 crc kubenswrapper[4888]: I1124 01:17:09.245903 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:17:09 crc kubenswrapper[4888]: E1124 01:17:09.246998 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:17:22 crc kubenswrapper[4888]: I1124 01:17:22.245619 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:17:22 crc kubenswrapper[4888]: E1124 01:17:22.246705 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:17:35 crc kubenswrapper[4888]: I1124 01:17:35.245258 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:17:35 crc kubenswrapper[4888]: E1124 01:17:35.246061 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:17:45 crc kubenswrapper[4888]: I1124 01:17:45.379681 4888 generic.go:334] "Generic (PLEG): container finished" podID="ebbc9282-a85b-4484-b130-5b0c91030009" containerID="389c0e903db9b9cff159f36e8c48c9bfaa6a78d97e49ee6fab54352f30683e2c" exitCode=0 Nov 24 01:17:45 crc kubenswrapper[4888]: I1124 01:17:45.379798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" event={"ID":"ebbc9282-a85b-4484-b130-5b0c91030009","Type":"ContainerDied","Data":"389c0e903db9b9cff159f36e8c48c9bfaa6a78d97e49ee6fab54352f30683e2c"} Nov 24 01:17:46 crc kubenswrapper[4888]: I1124 01:17:46.902143 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.025712 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key\") pod \"ebbc9282-a85b-4484-b130-5b0c91030009\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.026309 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv4tp\" (UniqueName: \"kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp\") pod \"ebbc9282-a85b-4484-b130-5b0c91030009\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.026415 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph\") pod \"ebbc9282-a85b-4484-b130-5b0c91030009\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.026474 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle\") pod \"ebbc9282-a85b-4484-b130-5b0c91030009\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.026535 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory\") pod \"ebbc9282-a85b-4484-b130-5b0c91030009\" (UID: \"ebbc9282-a85b-4484-b130-5b0c91030009\") " Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.032782 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph" (OuterVolumeSpecName: "ceph") pod "ebbc9282-a85b-4484-b130-5b0c91030009" (UID: "ebbc9282-a85b-4484-b130-5b0c91030009"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.032911 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp" (OuterVolumeSpecName: "kube-api-access-rv4tp") pod "ebbc9282-a85b-4484-b130-5b0c91030009" (UID: "ebbc9282-a85b-4484-b130-5b0c91030009"). InnerVolumeSpecName "kube-api-access-rv4tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.033606 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ebbc9282-a85b-4484-b130-5b0c91030009" (UID: "ebbc9282-a85b-4484-b130-5b0c91030009"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.067704 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory" (OuterVolumeSpecName: "inventory") pod "ebbc9282-a85b-4484-b130-5b0c91030009" (UID: "ebbc9282-a85b-4484-b130-5b0c91030009"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.067722 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ebbc9282-a85b-4484-b130-5b0c91030009" (UID: "ebbc9282-a85b-4484-b130-5b0c91030009"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.128910 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv4tp\" (UniqueName: \"kubernetes.io/projected/ebbc9282-a85b-4484-b130-5b0c91030009-kube-api-access-rv4tp\") on node \"crc\" DevicePath \"\"" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.128949 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.128963 4888 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.128975 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.128988 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebbc9282-a85b-4484-b130-5b0c91030009-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.245363 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:17:47 crc kubenswrapper[4888]: E1124 01:17:47.246219 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.409255 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" event={"ID":"ebbc9282-a85b-4484-b130-5b0c91030009","Type":"ContainerDied","Data":"d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3"} Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.409300 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6be22d5ae66b0964cc2f6e8171dee1a2c720cb563483e3d168e359f4a938ef3" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.409386 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.550679 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj"] Nov 24 01:17:47 crc kubenswrapper[4888]: E1124 01:17:47.551263 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="registry-server" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551286 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="registry-server" Nov 24 01:17:47 crc kubenswrapper[4888]: E1124 01:17:47.551313 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbc9282-a85b-4484-b130-5b0c91030009" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551322 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbc9282-a85b-4484-b130-5b0c91030009" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 01:17:47 crc kubenswrapper[4888]: E1124 01:17:47.551337 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="extract-utilities" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551345 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="extract-utilities" Nov 24 01:17:47 crc kubenswrapper[4888]: E1124 01:17:47.551376 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="extract-content" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551386 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="extract-content" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551637 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebbc9282-a85b-4484-b130-5b0c91030009" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.551875 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c310cbf-a058-4590-bd21-45eea8198d35" containerName="registry-server" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.553150 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.555894 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.556444 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.556591 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.556838 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.558356 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.577464 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj"] Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.640855 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95mnp\" (UniqueName: \"kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.640920 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.640947 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.641220 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.744082 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.744216 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95mnp\" (UniqueName: \"kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.744282 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.744317 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.750016 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.750494 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.750943 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.769778 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95mnp\" (UniqueName: \"kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r99nj\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:47 crc kubenswrapper[4888]: I1124 01:17:47.890551 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:17:48 crc kubenswrapper[4888]: I1124 01:17:48.370998 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj"] Nov 24 01:17:48 crc kubenswrapper[4888]: I1124 01:17:48.419905 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" event={"ID":"5f1ab899-c7e5-4618-b282-8eebc0dec80b","Type":"ContainerStarted","Data":"ca26e4aade904a63bff35319691317c89691defae1d7c82f3a0f4699e141fb7e"} Nov 24 01:17:49 crc kubenswrapper[4888]: I1124 01:17:49.433466 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" event={"ID":"5f1ab899-c7e5-4618-b282-8eebc0dec80b","Type":"ContainerStarted","Data":"e24ff795e05237c8edfe1331f46118a26bedbaf73c5cfad2a8c68d3d25463ce8"} Nov 24 01:17:49 crc kubenswrapper[4888]: I1124 01:17:49.461856 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" podStartSLOduration=1.9821106849999999 podStartE2EDuration="2.461803913s" podCreationTimestamp="2025-11-24 01:17:47 +0000 UTC" firstStartedPulling="2025-11-24 01:17:48.374045395 +0000 UTC m=+3170.956729439" lastFinishedPulling="2025-11-24 01:17:48.853738623 +0000 UTC m=+3171.436422667" observedRunningTime="2025-11-24 01:17:49.453264623 +0000 UTC m=+3172.035948687" watchObservedRunningTime="2025-11-24 01:17:49.461803913 +0000 UTC m=+3172.044487967" Nov 24 01:17:58 crc kubenswrapper[4888]: I1124 01:17:58.256278 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:17:58 crc kubenswrapper[4888]: I1124 01:17:58.583680 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d"} Nov 24 01:18:20 crc kubenswrapper[4888]: I1124 01:18:20.836115 4888 generic.go:334] "Generic (PLEG): container finished" podID="5f1ab899-c7e5-4618-b282-8eebc0dec80b" containerID="e24ff795e05237c8edfe1331f46118a26bedbaf73c5cfad2a8c68d3d25463ce8" exitCode=0 Nov 24 01:18:20 crc kubenswrapper[4888]: I1124 01:18:20.836230 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" event={"ID":"5f1ab899-c7e5-4618-b282-8eebc0dec80b","Type":"ContainerDied","Data":"e24ff795e05237c8edfe1331f46118a26bedbaf73c5cfad2a8c68d3d25463ce8"} Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.435558 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.452169 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key\") pod \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.452218 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph\") pod \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.452513 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95mnp\" (UniqueName: \"kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp\") pod \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.452552 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory\") pod \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\" (UID: \"5f1ab899-c7e5-4618-b282-8eebc0dec80b\") " Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.459439 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph" (OuterVolumeSpecName: "ceph") pod "5f1ab899-c7e5-4618-b282-8eebc0dec80b" (UID: "5f1ab899-c7e5-4618-b282-8eebc0dec80b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.463169 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp" (OuterVolumeSpecName: "kube-api-access-95mnp") pod "5f1ab899-c7e5-4618-b282-8eebc0dec80b" (UID: "5f1ab899-c7e5-4618-b282-8eebc0dec80b"). InnerVolumeSpecName "kube-api-access-95mnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.495957 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory" (OuterVolumeSpecName: "inventory") pod "5f1ab899-c7e5-4618-b282-8eebc0dec80b" (UID: "5f1ab899-c7e5-4618-b282-8eebc0dec80b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.502965 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f1ab899-c7e5-4618-b282-8eebc0dec80b" (UID: "5f1ab899-c7e5-4618-b282-8eebc0dec80b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.555208 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95mnp\" (UniqueName: \"kubernetes.io/projected/5f1ab899-c7e5-4618-b282-8eebc0dec80b-kube-api-access-95mnp\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.555266 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.555276 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.555284 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f1ab899-c7e5-4618-b282-8eebc0dec80b-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.871773 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" event={"ID":"5f1ab899-c7e5-4618-b282-8eebc0dec80b","Type":"ContainerDied","Data":"ca26e4aade904a63bff35319691317c89691defae1d7c82f3a0f4699e141fb7e"} Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.872092 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca26e4aade904a63bff35319691317c89691defae1d7c82f3a0f4699e141fb7e" Nov 24 01:18:22 crc kubenswrapper[4888]: I1124 01:18:22.871957 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r99nj" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.009026 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl"] Nov 24 01:18:23 crc kubenswrapper[4888]: E1124 01:18:23.011005 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1ab899-c7e5-4618-b282-8eebc0dec80b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.011049 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1ab899-c7e5-4618-b282-8eebc0dec80b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.011507 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f1ab899-c7e5-4618-b282-8eebc0dec80b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.012576 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.016796 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.019077 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.019339 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.019383 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.019628 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.021188 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl"] Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.066530 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.066728 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.066799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.066866 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45hc9\" (UniqueName: \"kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.168831 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.169281 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.169410 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.169461 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45hc9\" (UniqueName: \"kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.175547 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.175711 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.176555 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.197990 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45hc9\" (UniqueName: \"kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crwxl\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:23 crc kubenswrapper[4888]: I1124 01:18:23.344806 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:24 crc kubenswrapper[4888]: I1124 01:18:24.028773 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl"] Nov 24 01:18:24 crc kubenswrapper[4888]: I1124 01:18:24.898716 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" event={"ID":"de6be747-eec3-445b-a6e5-3cd034b68c6a","Type":"ContainerStarted","Data":"c92aa2a11a0ad019861e7d16279bdf94f051221713397c2bee1dcb67ff034da8"} Nov 24 01:18:25 crc kubenswrapper[4888]: I1124 01:18:25.909494 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" event={"ID":"de6be747-eec3-445b-a6e5-3cd034b68c6a","Type":"ContainerStarted","Data":"6f89a168ec9cc82cceec371b92fa459e2515ef1687b2fcc22b4c543718b95d82"} Nov 24 01:18:25 crc kubenswrapper[4888]: I1124 01:18:25.933259 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" podStartSLOduration=3.309984229 podStartE2EDuration="3.933240927s" podCreationTimestamp="2025-11-24 01:18:22 +0000 UTC" firstStartedPulling="2025-11-24 01:18:24.035801427 +0000 UTC m=+3206.618485471" lastFinishedPulling="2025-11-24 01:18:24.659058105 +0000 UTC m=+3207.241742169" observedRunningTime="2025-11-24 01:18:25.929693758 +0000 UTC m=+3208.512377812" watchObservedRunningTime="2025-11-24 01:18:25.933240927 +0000 UTC m=+3208.515924971" Nov 24 01:18:31 crc kubenswrapper[4888]: I1124 01:18:31.973032 4888 generic.go:334] "Generic (PLEG): container finished" podID="de6be747-eec3-445b-a6e5-3cd034b68c6a" containerID="6f89a168ec9cc82cceec371b92fa459e2515ef1687b2fcc22b4c543718b95d82" exitCode=0 Nov 24 01:18:31 crc kubenswrapper[4888]: I1124 01:18:31.973081 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" event={"ID":"de6be747-eec3-445b-a6e5-3cd034b68c6a","Type":"ContainerDied","Data":"6f89a168ec9cc82cceec371b92fa459e2515ef1687b2fcc22b4c543718b95d82"} Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.556220 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.634722 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45hc9\" (UniqueName: \"kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9\") pod \"de6be747-eec3-445b-a6e5-3cd034b68c6a\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.634934 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory\") pod \"de6be747-eec3-445b-a6e5-3cd034b68c6a\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.635168 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key\") pod \"de6be747-eec3-445b-a6e5-3cd034b68c6a\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.635291 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph\") pod \"de6be747-eec3-445b-a6e5-3cd034b68c6a\" (UID: \"de6be747-eec3-445b-a6e5-3cd034b68c6a\") " Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.640632 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9" (OuterVolumeSpecName: "kube-api-access-45hc9") pod "de6be747-eec3-445b-a6e5-3cd034b68c6a" (UID: "de6be747-eec3-445b-a6e5-3cd034b68c6a"). InnerVolumeSpecName "kube-api-access-45hc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.642196 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph" (OuterVolumeSpecName: "ceph") pod "de6be747-eec3-445b-a6e5-3cd034b68c6a" (UID: "de6be747-eec3-445b-a6e5-3cd034b68c6a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.671027 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de6be747-eec3-445b-a6e5-3cd034b68c6a" (UID: "de6be747-eec3-445b-a6e5-3cd034b68c6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.672939 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory" (OuterVolumeSpecName: "inventory") pod "de6be747-eec3-445b-a6e5-3cd034b68c6a" (UID: "de6be747-eec3-445b-a6e5-3cd034b68c6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.740114 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.740165 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.740188 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45hc9\" (UniqueName: \"kubernetes.io/projected/de6be747-eec3-445b-a6e5-3cd034b68c6a-kube-api-access-45hc9\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.740203 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de6be747-eec3-445b-a6e5-3cd034b68c6a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.998009 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" event={"ID":"de6be747-eec3-445b-a6e5-3cd034b68c6a","Type":"ContainerDied","Data":"c92aa2a11a0ad019861e7d16279bdf94f051221713397c2bee1dcb67ff034da8"} Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.998087 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c92aa2a11a0ad019861e7d16279bdf94f051221713397c2bee1dcb67ff034da8" Nov 24 01:18:33 crc kubenswrapper[4888]: I1124 01:18:33.998124 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crwxl" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.100171 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz"] Nov 24 01:18:34 crc kubenswrapper[4888]: E1124 01:18:34.101353 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6be747-eec3-445b-a6e5-3cd034b68c6a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.101385 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6be747-eec3-445b-a6e5-3cd034b68c6a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.101794 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6be747-eec3-445b-a6e5-3cd034b68c6a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.103163 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.107991 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.108243 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.108006 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.108644 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.114536 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.130519 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz"] Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.149913 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.149971 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.150064 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.150189 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.252080 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.252173 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.252297 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.252429 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.258634 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.258645 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.271327 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.278462 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rkgqz\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:34 crc kubenswrapper[4888]: I1124 01:18:34.435314 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:18:35 crc kubenswrapper[4888]: I1124 01:18:35.029896 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz"] Nov 24 01:18:36 crc kubenswrapper[4888]: I1124 01:18:36.020110 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" event={"ID":"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25","Type":"ContainerStarted","Data":"bd10cd8c0e672c61240e67bb5ef4052664d69b82af7609cc12487b0bd1456fbd"} Nov 24 01:18:36 crc kubenswrapper[4888]: I1124 01:18:36.020460 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" event={"ID":"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25","Type":"ContainerStarted","Data":"779bd3a2354a56859bfef1c2e89ff8b37245f7ce78afde869ef5a3ebe7c7b398"} Nov 24 01:18:36 crc kubenswrapper[4888]: I1124 01:18:36.056033 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" podStartSLOduration=1.5722444740000001 podStartE2EDuration="2.056009246s" podCreationTimestamp="2025-11-24 01:18:34 +0000 UTC" firstStartedPulling="2025-11-24 01:18:35.04257476 +0000 UTC m=+3217.625258804" lastFinishedPulling="2025-11-24 01:18:35.526339492 +0000 UTC m=+3218.109023576" observedRunningTime="2025-11-24 01:18:36.048420423 +0000 UTC m=+3218.631104467" watchObservedRunningTime="2025-11-24 01:18:36.056009246 +0000 UTC m=+3218.638693290" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.345502 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.349067 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.362048 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.531409 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.531530 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7sg2\" (UniqueName: \"kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.531612 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.634127 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7sg2\" (UniqueName: \"kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.634463 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.634753 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.635027 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.635296 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.657424 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7sg2\" (UniqueName: \"kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2\") pod \"certified-operators-rth99\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:41 crc kubenswrapper[4888]: I1124 01:18:41.695523 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:42 crc kubenswrapper[4888]: I1124 01:18:42.163025 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:43 crc kubenswrapper[4888]: I1124 01:18:43.103790 4888 generic.go:334] "Generic (PLEG): container finished" podID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerID="ec95381f28ac1a8963145df04186747df58234f2cab9256cfd1e50c09e10a4fd" exitCode=0 Nov 24 01:18:43 crc kubenswrapper[4888]: I1124 01:18:43.103863 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerDied","Data":"ec95381f28ac1a8963145df04186747df58234f2cab9256cfd1e50c09e10a4fd"} Nov 24 01:18:43 crc kubenswrapper[4888]: I1124 01:18:43.104147 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerStarted","Data":"d748bada5d4abc03490026ae945c77cf1a2e0afe7eed528db3b20a2a30eb0c2d"} Nov 24 01:18:44 crc kubenswrapper[4888]: I1124 01:18:44.122161 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerStarted","Data":"435e15b253f9bd4dd87501abfa2d82abe5fbec82234beae2d49865ba59acc775"} Nov 24 01:18:45 crc kubenswrapper[4888]: I1124 01:18:45.140770 4888 generic.go:334] "Generic (PLEG): container finished" podID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerID="435e15b253f9bd4dd87501abfa2d82abe5fbec82234beae2d49865ba59acc775" exitCode=0 Nov 24 01:18:45 crc kubenswrapper[4888]: I1124 01:18:45.141136 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerDied","Data":"435e15b253f9bd4dd87501abfa2d82abe5fbec82234beae2d49865ba59acc775"} Nov 24 01:18:46 crc kubenswrapper[4888]: I1124 01:18:46.155659 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerStarted","Data":"dd3261c101da221faf72a8c9044abe4af380ab78c079fca39c0a17f8025a280b"} Nov 24 01:18:46 crc kubenswrapper[4888]: I1124 01:18:46.181886 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rth99" podStartSLOduration=2.774246615 podStartE2EDuration="5.181861471s" podCreationTimestamp="2025-11-24 01:18:41 +0000 UTC" firstStartedPulling="2025-11-24 01:18:43.106941787 +0000 UTC m=+3225.689625841" lastFinishedPulling="2025-11-24 01:18:45.514556613 +0000 UTC m=+3228.097240697" observedRunningTime="2025-11-24 01:18:46.174634807 +0000 UTC m=+3228.757318871" watchObservedRunningTime="2025-11-24 01:18:46.181861471 +0000 UTC m=+3228.764545575" Nov 24 01:18:51 crc kubenswrapper[4888]: I1124 01:18:51.697296 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:51 crc kubenswrapper[4888]: I1124 01:18:51.698088 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:51 crc kubenswrapper[4888]: I1124 01:18:51.802254 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:52 crc kubenswrapper[4888]: I1124 01:18:52.299763 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:52 crc kubenswrapper[4888]: I1124 01:18:52.366503 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:54 crc kubenswrapper[4888]: I1124 01:18:54.252001 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rth99" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="registry-server" containerID="cri-o://dd3261c101da221faf72a8c9044abe4af380ab78c079fca39c0a17f8025a280b" gracePeriod=2 Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.264195 4888 generic.go:334] "Generic (PLEG): container finished" podID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerID="dd3261c101da221faf72a8c9044abe4af380ab78c079fca39c0a17f8025a280b" exitCode=0 Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.264264 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerDied","Data":"dd3261c101da221faf72a8c9044abe4af380ab78c079fca39c0a17f8025a280b"} Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.264621 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rth99" event={"ID":"96bc14fc-b701-406e-b29d-6d95a0867f39","Type":"ContainerDied","Data":"d748bada5d4abc03490026ae945c77cf1a2e0afe7eed528db3b20a2a30eb0c2d"} Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.264645 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d748bada5d4abc03490026ae945c77cf1a2e0afe7eed528db3b20a2a30eb0c2d" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.339941 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.477346 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7sg2\" (UniqueName: \"kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2\") pod \"96bc14fc-b701-406e-b29d-6d95a0867f39\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.477475 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content\") pod \"96bc14fc-b701-406e-b29d-6d95a0867f39\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.477659 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities\") pod \"96bc14fc-b701-406e-b29d-6d95a0867f39\" (UID: \"96bc14fc-b701-406e-b29d-6d95a0867f39\") " Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.479641 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities" (OuterVolumeSpecName: "utilities") pod "96bc14fc-b701-406e-b29d-6d95a0867f39" (UID: "96bc14fc-b701-406e-b29d-6d95a0867f39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.483888 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2" (OuterVolumeSpecName: "kube-api-access-l7sg2") pod "96bc14fc-b701-406e-b29d-6d95a0867f39" (UID: "96bc14fc-b701-406e-b29d-6d95a0867f39"). InnerVolumeSpecName "kube-api-access-l7sg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.543495 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96bc14fc-b701-406e-b29d-6d95a0867f39" (UID: "96bc14fc-b701-406e-b29d-6d95a0867f39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.580698 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.580727 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96bc14fc-b701-406e-b29d-6d95a0867f39-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:55 crc kubenswrapper[4888]: I1124 01:18:55.580738 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7sg2\" (UniqueName: \"kubernetes.io/projected/96bc14fc-b701-406e-b29d-6d95a0867f39-kube-api-access-l7sg2\") on node \"crc\" DevicePath \"\"" Nov 24 01:18:56 crc kubenswrapper[4888]: I1124 01:18:56.274938 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rth99" Nov 24 01:18:56 crc kubenswrapper[4888]: I1124 01:18:56.329757 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:56 crc kubenswrapper[4888]: I1124 01:18:56.341689 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rth99"] Nov 24 01:18:58 crc kubenswrapper[4888]: I1124 01:18:58.276041 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" path="/var/lib/kubelet/pods/96bc14fc-b701-406e-b29d-6d95a0867f39/volumes" Nov 24 01:19:20 crc kubenswrapper[4888]: I1124 01:19:20.560633 4888 generic.go:334] "Generic (PLEG): container finished" podID="e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" containerID="bd10cd8c0e672c61240e67bb5ef4052664d69b82af7609cc12487b0bd1456fbd" exitCode=0 Nov 24 01:19:20 crc kubenswrapper[4888]: I1124 01:19:20.560709 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" event={"ID":"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25","Type":"ContainerDied","Data":"bd10cd8c0e672c61240e67bb5ef4052664d69b82af7609cc12487b0bd1456fbd"} Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.034513 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.208873 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph\") pod \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.209140 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory\") pod \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.209224 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd\") pod \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.209263 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key\") pod \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\" (UID: \"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25\") " Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.214205 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd" (OuterVolumeSpecName: "kube-api-access-x5xkd") pod "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" (UID: "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25"). InnerVolumeSpecName "kube-api-access-x5xkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.215429 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph" (OuterVolumeSpecName: "ceph") pod "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" (UID: "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.250519 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" (UID: "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.266202 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory" (OuterVolumeSpecName: "inventory") pod "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" (UID: "e5f0ab5f-207b-4e2b-826a-87daf6aa5f25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.312982 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.313053 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.313081 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-kube-api-access-x5xkd\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.313104 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f0ab5f-207b-4e2b-826a-87daf6aa5f25-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.586487 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" event={"ID":"e5f0ab5f-207b-4e2b-826a-87daf6aa5f25","Type":"ContainerDied","Data":"779bd3a2354a56859bfef1c2e89ff8b37245f7ce78afde869ef5a3ebe7c7b398"} Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.586541 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="779bd3a2354a56859bfef1c2e89ff8b37245f7ce78afde869ef5a3ebe7c7b398" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.586602 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rkgqz" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.684866 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds"] Nov 24 01:19:22 crc kubenswrapper[4888]: E1124 01:19:22.686055 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.686151 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:22 crc kubenswrapper[4888]: E1124 01:19:22.686252 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="extract-utilities" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.686323 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="extract-utilities" Nov 24 01:19:22 crc kubenswrapper[4888]: E1124 01:19:22.686440 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="registry-server" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.686529 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="registry-server" Nov 24 01:19:22 crc kubenswrapper[4888]: E1124 01:19:22.686612 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="extract-content" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.686679 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="extract-content" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.687044 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f0ab5f-207b-4e2b-826a-87daf6aa5f25" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.687154 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="96bc14fc-b701-406e-b29d-6d95a0867f39" containerName="registry-server" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.688164 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.690828 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.691350 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.691583 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.691787 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.692194 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.701657 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds"] Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.824400 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfhns\" (UniqueName: \"kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.824442 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.824481 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.824501 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.927379 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfhns\" (UniqueName: \"kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.927445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.927499 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.927525 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.933501 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.933508 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.934149 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:22 crc kubenswrapper[4888]: I1124 01:19:22.943748 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfhns\" (UniqueName: \"kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:23 crc kubenswrapper[4888]: I1124 01:19:23.013910 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:23 crc kubenswrapper[4888]: I1124 01:19:23.593640 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds"] Nov 24 01:19:24 crc kubenswrapper[4888]: I1124 01:19:24.607721 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" event={"ID":"99022481-7ff1-4b5a-a06e-376fbeaa29ba","Type":"ContainerStarted","Data":"4b1d6bdc40b3238306772bfef3e1a5e4d123c35307e04b1e64ded2a4eb984f5a"} Nov 24 01:19:24 crc kubenswrapper[4888]: I1124 01:19:24.608351 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" event={"ID":"99022481-7ff1-4b5a-a06e-376fbeaa29ba","Type":"ContainerStarted","Data":"56a40562d5c902256a1f2c0d9fb79856b0e7058d9796e05bacc9a3a6afc0be70"} Nov 24 01:19:24 crc kubenswrapper[4888]: I1124 01:19:24.634962 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" podStartSLOduration=2.158977632 podStartE2EDuration="2.634943975s" podCreationTimestamp="2025-11-24 01:19:22 +0000 UTC" firstStartedPulling="2025-11-24 01:19:23.60060598 +0000 UTC m=+3266.183290034" lastFinishedPulling="2025-11-24 01:19:24.076572323 +0000 UTC m=+3266.659256377" observedRunningTime="2025-11-24 01:19:24.627985039 +0000 UTC m=+3267.210669093" watchObservedRunningTime="2025-11-24 01:19:24.634943975 +0000 UTC m=+3267.217628019" Nov 24 01:19:28 crc kubenswrapper[4888]: I1124 01:19:28.658273 4888 generic.go:334] "Generic (PLEG): container finished" podID="99022481-7ff1-4b5a-a06e-376fbeaa29ba" containerID="4b1d6bdc40b3238306772bfef3e1a5e4d123c35307e04b1e64ded2a4eb984f5a" exitCode=0 Nov 24 01:19:28 crc kubenswrapper[4888]: I1124 01:19:28.658526 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" event={"ID":"99022481-7ff1-4b5a-a06e-376fbeaa29ba","Type":"ContainerDied","Data":"4b1d6bdc40b3238306772bfef3e1a5e4d123c35307e04b1e64ded2a4eb984f5a"} Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.126436 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.303374 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph\") pod \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.303505 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory\") pod \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.303595 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key\") pod \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.303727 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfhns\" (UniqueName: \"kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns\") pod \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\" (UID: \"99022481-7ff1-4b5a-a06e-376fbeaa29ba\") " Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.308634 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns" (OuterVolumeSpecName: "kube-api-access-kfhns") pod "99022481-7ff1-4b5a-a06e-376fbeaa29ba" (UID: "99022481-7ff1-4b5a-a06e-376fbeaa29ba"). InnerVolumeSpecName "kube-api-access-kfhns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.310282 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph" (OuterVolumeSpecName: "ceph") pod "99022481-7ff1-4b5a-a06e-376fbeaa29ba" (UID: "99022481-7ff1-4b5a-a06e-376fbeaa29ba"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.332998 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99022481-7ff1-4b5a-a06e-376fbeaa29ba" (UID: "99022481-7ff1-4b5a-a06e-376fbeaa29ba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.334366 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory" (OuterVolumeSpecName: "inventory") pod "99022481-7ff1-4b5a-a06e-376fbeaa29ba" (UID: "99022481-7ff1-4b5a-a06e-376fbeaa29ba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.406498 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.406535 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfhns\" (UniqueName: \"kubernetes.io/projected/99022481-7ff1-4b5a-a06e-376fbeaa29ba-kube-api-access-kfhns\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.406546 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.406556 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99022481-7ff1-4b5a-a06e-376fbeaa29ba-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.686238 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.686148 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds" event={"ID":"99022481-7ff1-4b5a-a06e-376fbeaa29ba","Type":"ContainerDied","Data":"56a40562d5c902256a1f2c0d9fb79856b0e7058d9796e05bacc9a3a6afc0be70"} Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.687016 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56a40562d5c902256a1f2c0d9fb79856b0e7058d9796e05bacc9a3a6afc0be70" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.768363 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr"] Nov 24 01:19:30 crc kubenswrapper[4888]: E1124 01:19:30.769113 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99022481-7ff1-4b5a-a06e-376fbeaa29ba" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.769143 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="99022481-7ff1-4b5a-a06e-376fbeaa29ba" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.769548 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="99022481-7ff1-4b5a-a06e-376fbeaa29ba" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.770787 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.773802 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.774228 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.774408 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.774451 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.778618 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.782107 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr"] Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.918805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xswp8\" (UniqueName: \"kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.918939 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.919167 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:30 crc kubenswrapper[4888]: I1124 01:19:30.919291 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.021359 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xswp8\" (UniqueName: \"kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.021416 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.021454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.021491 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.025894 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.026129 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.027312 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.052952 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xswp8\" (UniqueName: \"kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.099549 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:19:31 crc kubenswrapper[4888]: I1124 01:19:31.733391 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr"] Nov 24 01:19:32 crc kubenswrapper[4888]: I1124 01:19:32.706730 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" event={"ID":"32bcbd1c-35f4-4632-ab42-01e6cc578d93","Type":"ContainerStarted","Data":"b18f36d1bc7d2a64e53106fb9898f4896993f7af8a67b521065d7e7f7164f1dc"} Nov 24 01:19:32 crc kubenswrapper[4888]: I1124 01:19:32.707135 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" event={"ID":"32bcbd1c-35f4-4632-ab42-01e6cc578d93","Type":"ContainerStarted","Data":"3dfd66660800224a185a8d000a979db02bb8f186c97ee2e6a09b2e438db6d126"} Nov 24 01:19:32 crc kubenswrapper[4888]: I1124 01:19:32.732277 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" podStartSLOduration=2.299272718 podStartE2EDuration="2.732254771s" podCreationTimestamp="2025-11-24 01:19:30 +0000 UTC" firstStartedPulling="2025-11-24 01:19:31.735480553 +0000 UTC m=+3274.318164597" lastFinishedPulling="2025-11-24 01:19:32.168462566 +0000 UTC m=+3274.751146650" observedRunningTime="2025-11-24 01:19:32.729702589 +0000 UTC m=+3275.312386643" watchObservedRunningTime="2025-11-24 01:19:32.732254771 +0000 UTC m=+3275.314938825" Nov 24 01:20:23 crc kubenswrapper[4888]: I1124 01:20:23.591981 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:20:23 crc kubenswrapper[4888]: I1124 01:20:23.592650 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:20:28 crc kubenswrapper[4888]: I1124 01:20:28.883475 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:28 crc kubenswrapper[4888]: I1124 01:20:28.886417 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:28 crc kubenswrapper[4888]: I1124 01:20:28.893584 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.000111 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.000269 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqr67\" (UniqueName: \"kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.000625 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.104039 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.104104 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqr67\" (UniqueName: \"kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.104429 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.104667 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.105107 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.133027 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqr67\" (UniqueName: \"kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67\") pod \"redhat-marketplace-mmwqs\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.208160 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:29 crc kubenswrapper[4888]: I1124 01:20:29.730507 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:30 crc kubenswrapper[4888]: I1124 01:20:30.352700 4888 generic.go:334] "Generic (PLEG): container finished" podID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerID="27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8" exitCode=0 Nov 24 01:20:30 crc kubenswrapper[4888]: I1124 01:20:30.352795 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerDied","Data":"27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8"} Nov 24 01:20:30 crc kubenswrapper[4888]: I1124 01:20:30.353068 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerStarted","Data":"f94b768c939befdb97b556533798de021d8e1654a0889037cfc7f18080937743"} Nov 24 01:20:30 crc kubenswrapper[4888]: I1124 01:20:30.355469 4888 generic.go:334] "Generic (PLEG): container finished" podID="32bcbd1c-35f4-4632-ab42-01e6cc578d93" containerID="b18f36d1bc7d2a64e53106fb9898f4896993f7af8a67b521065d7e7f7164f1dc" exitCode=0 Nov 24 01:20:30 crc kubenswrapper[4888]: I1124 01:20:30.355519 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" event={"ID":"32bcbd1c-35f4-4632-ab42-01e6cc578d93","Type":"ContainerDied","Data":"b18f36d1bc7d2a64e53106fb9898f4896993f7af8a67b521065d7e7f7164f1dc"} Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.369006 4888 generic.go:334] "Generic (PLEG): container finished" podID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerID="1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8" exitCode=0 Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.369103 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerDied","Data":"1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8"} Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.824825 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.873453 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory\") pod \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.873775 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key\") pod \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.873902 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xswp8\" (UniqueName: \"kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8\") pod \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.874033 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph\") pod \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\" (UID: \"32bcbd1c-35f4-4632-ab42-01e6cc578d93\") " Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.878957 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph" (OuterVolumeSpecName: "ceph") pod "32bcbd1c-35f4-4632-ab42-01e6cc578d93" (UID: "32bcbd1c-35f4-4632-ab42-01e6cc578d93"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.879242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8" (OuterVolumeSpecName: "kube-api-access-xswp8") pod "32bcbd1c-35f4-4632-ab42-01e6cc578d93" (UID: "32bcbd1c-35f4-4632-ab42-01e6cc578d93"). InnerVolumeSpecName "kube-api-access-xswp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.913291 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32bcbd1c-35f4-4632-ab42-01e6cc578d93" (UID: "32bcbd1c-35f4-4632-ab42-01e6cc578d93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.919068 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory" (OuterVolumeSpecName: "inventory") pod "32bcbd1c-35f4-4632-ab42-01e6cc578d93" (UID: "32bcbd1c-35f4-4632-ab42-01e6cc578d93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.976387 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.976418 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.976427 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xswp8\" (UniqueName: \"kubernetes.io/projected/32bcbd1c-35f4-4632-ab42-01e6cc578d93-kube-api-access-xswp8\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:31 crc kubenswrapper[4888]: I1124 01:20:31.976437 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32bcbd1c-35f4-4632-ab42-01e6cc578d93-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.383496 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" event={"ID":"32bcbd1c-35f4-4632-ab42-01e6cc578d93","Type":"ContainerDied","Data":"3dfd66660800224a185a8d000a979db02bb8f186c97ee2e6a09b2e438db6d126"} Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.383549 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dfd66660800224a185a8d000a979db02bb8f186c97ee2e6a09b2e438db6d126" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.383565 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.388548 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerStarted","Data":"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4"} Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.413896 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mmwqs" podStartSLOduration=2.971394823 podStartE2EDuration="4.413873611s" podCreationTimestamp="2025-11-24 01:20:28 +0000 UTC" firstStartedPulling="2025-11-24 01:20:30.356323185 +0000 UTC m=+3332.939007269" lastFinishedPulling="2025-11-24 01:20:31.798802013 +0000 UTC m=+3334.381486057" observedRunningTime="2025-11-24 01:20:32.413188271 +0000 UTC m=+3334.995872315" watchObservedRunningTime="2025-11-24 01:20:32.413873611 +0000 UTC m=+3334.996557675" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.484605 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pbl78"] Nov 24 01:20:32 crc kubenswrapper[4888]: E1124 01:20:32.485690 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32bcbd1c-35f4-4632-ab42-01e6cc578d93" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.485718 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="32bcbd1c-35f4-4632-ab42-01e6cc578d93" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.485995 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="32bcbd1c-35f4-4632-ab42-01e6cc578d93" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.486981 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.489687 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.490533 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.490693 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.491665 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.494032 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.501739 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pbl78"] Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.590252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.590589 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.590725 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnn7t\" (UniqueName: \"kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.591061 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.692792 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.692895 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnn7t\" (UniqueName: \"kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.693023 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.693082 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.700889 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.701430 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.701469 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.709031 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnn7t\" (UniqueName: \"kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t\") pod \"ssh-known-hosts-edpm-deployment-pbl78\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:32 crc kubenswrapper[4888]: I1124 01:20:32.813949 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:33 crc kubenswrapper[4888]: I1124 01:20:33.360543 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pbl78"] Nov 24 01:20:33 crc kubenswrapper[4888]: W1124 01:20:33.361666 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8804f688_d740_4f33_ba57_ac0ca3d63a2b.slice/crio-3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9 WatchSource:0}: Error finding container 3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9: Status 404 returned error can't find the container with id 3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9 Nov 24 01:20:33 crc kubenswrapper[4888]: I1124 01:20:33.407311 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" event={"ID":"8804f688-d740-4f33-ba57-ac0ca3d63a2b","Type":"ContainerStarted","Data":"3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9"} Nov 24 01:20:34 crc kubenswrapper[4888]: I1124 01:20:34.423725 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" event={"ID":"8804f688-d740-4f33-ba57-ac0ca3d63a2b","Type":"ContainerStarted","Data":"0e906c001f9ea91a25cc71a60fd448322a0e8d64b921e905faa34306b494ca63"} Nov 24 01:20:34 crc kubenswrapper[4888]: I1124 01:20:34.446361 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" podStartSLOduration=1.998132048 podStartE2EDuration="2.44633495s" podCreationTimestamp="2025-11-24 01:20:32 +0000 UTC" firstStartedPulling="2025-11-24 01:20:33.364128629 +0000 UTC m=+3335.946812673" lastFinishedPulling="2025-11-24 01:20:33.812331491 +0000 UTC m=+3336.395015575" observedRunningTime="2025-11-24 01:20:34.445713833 +0000 UTC m=+3337.028397887" watchObservedRunningTime="2025-11-24 01:20:34.44633495 +0000 UTC m=+3337.029019004" Nov 24 01:20:39 crc kubenswrapper[4888]: I1124 01:20:39.209044 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:39 crc kubenswrapper[4888]: I1124 01:20:39.209915 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:39 crc kubenswrapper[4888]: I1124 01:20:39.296010 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:39 crc kubenswrapper[4888]: I1124 01:20:39.543078 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:39 crc kubenswrapper[4888]: I1124 01:20:39.593334 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:41 crc kubenswrapper[4888]: I1124 01:20:41.509611 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mmwqs" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="registry-server" containerID="cri-o://5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4" gracePeriod=2 Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.115931 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.210425 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities\") pod \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.210852 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content\") pod \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.210994 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqr67\" (UniqueName: \"kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67\") pod \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\" (UID: \"c9e4a511-8f06-4b0a-8315-4a2a3a189068\") " Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.211362 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities" (OuterVolumeSpecName: "utilities") pod "c9e4a511-8f06-4b0a-8315-4a2a3a189068" (UID: "c9e4a511-8f06-4b0a-8315-4a2a3a189068"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.212804 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.223519 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67" (OuterVolumeSpecName: "kube-api-access-cqr67") pod "c9e4a511-8f06-4b0a-8315-4a2a3a189068" (UID: "c9e4a511-8f06-4b0a-8315-4a2a3a189068"). InnerVolumeSpecName "kube-api-access-cqr67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.238981 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9e4a511-8f06-4b0a-8315-4a2a3a189068" (UID: "c9e4a511-8f06-4b0a-8315-4a2a3a189068"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.316540 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e4a511-8f06-4b0a-8315-4a2a3a189068-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.316582 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqr67\" (UniqueName: \"kubernetes.io/projected/c9e4a511-8f06-4b0a-8315-4a2a3a189068-kube-api-access-cqr67\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.524855 4888 generic.go:334] "Generic (PLEG): container finished" podID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerID="5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4" exitCode=0 Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.524947 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerDied","Data":"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4"} Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.524980 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmwqs" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.525073 4888 scope.go:117] "RemoveContainer" containerID="5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.525049 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmwqs" event={"ID":"c9e4a511-8f06-4b0a-8315-4a2a3a189068","Type":"ContainerDied","Data":"f94b768c939befdb97b556533798de021d8e1654a0889037cfc7f18080937743"} Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.548451 4888 scope.go:117] "RemoveContainer" containerID="1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.554685 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.563727 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmwqs"] Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.582901 4888 scope.go:117] "RemoveContainer" containerID="27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.649361 4888 scope.go:117] "RemoveContainer" containerID="5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4" Nov 24 01:20:42 crc kubenswrapper[4888]: E1124 01:20:42.650177 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4\": container with ID starting with 5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4 not found: ID does not exist" containerID="5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.650216 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4"} err="failed to get container status \"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4\": rpc error: code = NotFound desc = could not find container \"5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4\": container with ID starting with 5d34bfb08d44829b051187ab24f33a2af934832efb639d430882a023b65a5cc4 not found: ID does not exist" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.650243 4888 scope.go:117] "RemoveContainer" containerID="1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8" Nov 24 01:20:42 crc kubenswrapper[4888]: E1124 01:20:42.650542 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8\": container with ID starting with 1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8 not found: ID does not exist" containerID="1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.650608 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8"} err="failed to get container status \"1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8\": rpc error: code = NotFound desc = could not find container \"1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8\": container with ID starting with 1881ae5e7321ceed7bbe8fa0cf643cc464f1a002266304ab4a8fd994f065c6b8 not found: ID does not exist" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.650650 4888 scope.go:117] "RemoveContainer" containerID="27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8" Nov 24 01:20:42 crc kubenswrapper[4888]: E1124 01:20:42.651125 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8\": container with ID starting with 27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8 not found: ID does not exist" containerID="27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8" Nov 24 01:20:42 crc kubenswrapper[4888]: I1124 01:20:42.651164 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8"} err="failed to get container status \"27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8\": rpc error: code = NotFound desc = could not find container \"27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8\": container with ID starting with 27f91e3c08c1cf81e7a65be5ad7f380717649c6d4943023648fd6c3825161ae8 not found: ID does not exist" Nov 24 01:20:44 crc kubenswrapper[4888]: I1124 01:20:44.264065 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" path="/var/lib/kubelet/pods/c9e4a511-8f06-4b0a-8315-4a2a3a189068/volumes" Nov 24 01:20:46 crc kubenswrapper[4888]: I1124 01:20:46.570296 4888 generic.go:334] "Generic (PLEG): container finished" podID="8804f688-d740-4f33-ba57-ac0ca3d63a2b" containerID="0e906c001f9ea91a25cc71a60fd448322a0e8d64b921e905faa34306b494ca63" exitCode=0 Nov 24 01:20:46 crc kubenswrapper[4888]: I1124 01:20:46.570385 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" event={"ID":"8804f688-d740-4f33-ba57-ac0ca3d63a2b","Type":"ContainerDied","Data":"0e906c001f9ea91a25cc71a60fd448322a0e8d64b921e905faa34306b494ca63"} Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.134311 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.262614 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0\") pod \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.262686 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnn7t\" (UniqueName: \"kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t\") pod \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.262772 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam\") pod \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.262869 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph\") pod \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\" (UID: \"8804f688-d740-4f33-ba57-ac0ca3d63a2b\") " Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.269467 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t" (OuterVolumeSpecName: "kube-api-access-pnn7t") pod "8804f688-d740-4f33-ba57-ac0ca3d63a2b" (UID: "8804f688-d740-4f33-ba57-ac0ca3d63a2b"). InnerVolumeSpecName "kube-api-access-pnn7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.270086 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph" (OuterVolumeSpecName: "ceph") pod "8804f688-d740-4f33-ba57-ac0ca3d63a2b" (UID: "8804f688-d740-4f33-ba57-ac0ca3d63a2b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.300296 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8804f688-d740-4f33-ba57-ac0ca3d63a2b" (UID: "8804f688-d740-4f33-ba57-ac0ca3d63a2b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.313977 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "8804f688-d740-4f33-ba57-ac0ca3d63a2b" (UID: "8804f688-d740-4f33-ba57-ac0ca3d63a2b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.366061 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.366097 4888 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.366108 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnn7t\" (UniqueName: \"kubernetes.io/projected/8804f688-d740-4f33-ba57-ac0ca3d63a2b-kube-api-access-pnn7t\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.366117 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8804f688-d740-4f33-ba57-ac0ca3d63a2b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.598015 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.598108 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pbl78" event={"ID":"8804f688-d740-4f33-ba57-ac0ca3d63a2b","Type":"ContainerDied","Data":"3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9"} Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.598180 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3db867ddb02b406a236d98a42847131f3aa5104b0ca753589f41f28d021b5ea9" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.684362 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2"] Nov 24 01:20:48 crc kubenswrapper[4888]: E1124 01:20:48.685330 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="registry-server" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685361 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="registry-server" Nov 24 01:20:48 crc kubenswrapper[4888]: E1124 01:20:48.685395 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="extract-utilities" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685426 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="extract-utilities" Nov 24 01:20:48 crc kubenswrapper[4888]: E1124 01:20:48.685486 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8804f688-d740-4f33-ba57-ac0ca3d63a2b" containerName="ssh-known-hosts-edpm-deployment" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685499 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8804f688-d740-4f33-ba57-ac0ca3d63a2b" containerName="ssh-known-hosts-edpm-deployment" Nov 24 01:20:48 crc kubenswrapper[4888]: E1124 01:20:48.685523 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="extract-content" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685533 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="extract-content" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685857 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e4a511-8f06-4b0a-8315-4a2a3a189068" containerName="registry-server" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.685885 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8804f688-d740-4f33-ba57-ac0ca3d63a2b" containerName="ssh-known-hosts-edpm-deployment" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.686901 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.690188 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.690238 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.690281 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.690248 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.692284 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.709055 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2"] Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.875798 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.875881 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.875902 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqchx\" (UniqueName: \"kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.875953 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.978936 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.979207 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.979283 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqchx\" (UniqueName: \"kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.980120 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.985763 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.985924 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:48 crc kubenswrapper[4888]: I1124 01:20:48.986976 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:49 crc kubenswrapper[4888]: I1124 01:20:49.010587 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqchx\" (UniqueName: \"kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jrjs2\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:49 crc kubenswrapper[4888]: I1124 01:20:49.309134 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:20:49 crc kubenswrapper[4888]: I1124 01:20:49.974925 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2"] Nov 24 01:20:49 crc kubenswrapper[4888]: W1124 01:20:49.987774 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbab960db_54de_4def_b56e_9f31b8df0e03.slice/crio-d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87 WatchSource:0}: Error finding container d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87: Status 404 returned error can't find the container with id d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87 Nov 24 01:20:50 crc kubenswrapper[4888]: I1124 01:20:50.622647 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" event={"ID":"bab960db-54de-4def-b56e-9f31b8df0e03","Type":"ContainerStarted","Data":"d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87"} Nov 24 01:20:51 crc kubenswrapper[4888]: I1124 01:20:51.640016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" event={"ID":"bab960db-54de-4def-b56e-9f31b8df0e03","Type":"ContainerStarted","Data":"025d3639772be459d08843a762861f8040215e318a5018fe36ba28b2911ce586"} Nov 24 01:20:51 crc kubenswrapper[4888]: I1124 01:20:51.679447 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" podStartSLOduration=3.143141592 podStartE2EDuration="3.679426362s" podCreationTimestamp="2025-11-24 01:20:48 +0000 UTC" firstStartedPulling="2025-11-24 01:20:49.991036493 +0000 UTC m=+3352.573720567" lastFinishedPulling="2025-11-24 01:20:50.527321263 +0000 UTC m=+3353.110005337" observedRunningTime="2025-11-24 01:20:51.662752592 +0000 UTC m=+3354.245436646" watchObservedRunningTime="2025-11-24 01:20:51.679426362 +0000 UTC m=+3354.262110416" Nov 24 01:20:53 crc kubenswrapper[4888]: I1124 01:20:53.591564 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:20:53 crc kubenswrapper[4888]: I1124 01:20:53.592144 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:21:00 crc kubenswrapper[4888]: I1124 01:21:00.759100 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" event={"ID":"bab960db-54de-4def-b56e-9f31b8df0e03","Type":"ContainerDied","Data":"025d3639772be459d08843a762861f8040215e318a5018fe36ba28b2911ce586"} Nov 24 01:21:00 crc kubenswrapper[4888]: I1124 01:21:00.759052 4888 generic.go:334] "Generic (PLEG): container finished" podID="bab960db-54de-4def-b56e-9f31b8df0e03" containerID="025d3639772be459d08843a762861f8040215e318a5018fe36ba28b2911ce586" exitCode=0 Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.283781 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.413005 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqchx\" (UniqueName: \"kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx\") pod \"bab960db-54de-4def-b56e-9f31b8df0e03\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.413133 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key\") pod \"bab960db-54de-4def-b56e-9f31b8df0e03\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.413233 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph\") pod \"bab960db-54de-4def-b56e-9f31b8df0e03\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.413317 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory\") pod \"bab960db-54de-4def-b56e-9f31b8df0e03\" (UID: \"bab960db-54de-4def-b56e-9f31b8df0e03\") " Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.419585 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx" (OuterVolumeSpecName: "kube-api-access-sqchx") pod "bab960db-54de-4def-b56e-9f31b8df0e03" (UID: "bab960db-54de-4def-b56e-9f31b8df0e03"). InnerVolumeSpecName "kube-api-access-sqchx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.421022 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph" (OuterVolumeSpecName: "ceph") pod "bab960db-54de-4def-b56e-9f31b8df0e03" (UID: "bab960db-54de-4def-b56e-9f31b8df0e03"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.457241 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory" (OuterVolumeSpecName: "inventory") pod "bab960db-54de-4def-b56e-9f31b8df0e03" (UID: "bab960db-54de-4def-b56e-9f31b8df0e03"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.467530 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bab960db-54de-4def-b56e-9f31b8df0e03" (UID: "bab960db-54de-4def-b56e-9f31b8df0e03"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.516447 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqchx\" (UniqueName: \"kubernetes.io/projected/bab960db-54de-4def-b56e-9f31b8df0e03-kube-api-access-sqchx\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.516490 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.516509 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.516521 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bab960db-54de-4def-b56e-9f31b8df0e03-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.783512 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" event={"ID":"bab960db-54de-4def-b56e-9f31b8df0e03","Type":"ContainerDied","Data":"d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87"} Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.783574 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d51dab0f0acd8a4255879fc9e4a5e4dd70a94a45ff0c401f0c1f82c22bd7bf87" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.783580 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jrjs2" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.893288 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll"] Nov 24 01:21:02 crc kubenswrapper[4888]: E1124 01:21:02.893773 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab960db-54de-4def-b56e-9f31b8df0e03" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.893793 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab960db-54de-4def-b56e-9f31b8df0e03" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.894045 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab960db-54de-4def-b56e-9f31b8df0e03" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.894787 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.899026 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.899360 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.899500 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.900019 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.900157 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:21:02 crc kubenswrapper[4888]: I1124 01:21:02.910103 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll"] Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.026346 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.026393 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn9q7\" (UniqueName: \"kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.026653 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.026837 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.129216 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.129321 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn9q7\" (UniqueName: \"kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.129515 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.129607 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.136078 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.137049 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.137341 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.160697 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn9q7\" (UniqueName: \"kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.216470 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.630543 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll"] Nov 24 01:21:03 crc kubenswrapper[4888]: I1124 01:21:03.797542 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" event={"ID":"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797","Type":"ContainerStarted","Data":"dffd3df3131caa1ce2d69945f16927a5f3483a1b83a868268cf5cb3ff27d5d4d"} Nov 24 01:21:04 crc kubenswrapper[4888]: I1124 01:21:04.814173 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" event={"ID":"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797","Type":"ContainerStarted","Data":"70dfdaafb708638d19d2bca3c1c538cfac1420144d2bad56d489f780282c8a76"} Nov 24 01:21:15 crc kubenswrapper[4888]: I1124 01:21:15.972508 4888 generic.go:334] "Generic (PLEG): container finished" podID="2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" containerID="70dfdaafb708638d19d2bca3c1c538cfac1420144d2bad56d489f780282c8a76" exitCode=0 Nov 24 01:21:15 crc kubenswrapper[4888]: I1124 01:21:15.972630 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" event={"ID":"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797","Type":"ContainerDied","Data":"70dfdaafb708638d19d2bca3c1c538cfac1420144d2bad56d489f780282c8a76"} Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.544396 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.605555 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key\") pod \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.605636 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory\") pod \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.605769 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph\") pod \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.605995 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn9q7\" (UniqueName: \"kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7\") pod \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\" (UID: \"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797\") " Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.612205 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph" (OuterVolumeSpecName: "ceph") pod "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" (UID: "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.613314 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7" (OuterVolumeSpecName: "kube-api-access-cn9q7") pod "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" (UID: "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797"). InnerVolumeSpecName "kube-api-access-cn9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.645390 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" (UID: "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.655485 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory" (OuterVolumeSpecName: "inventory") pod "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" (UID: "2d25d7f9-02eb-4f78-a3c9-2f9a9a533797"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.710463 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.710511 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.710529 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:17 crc kubenswrapper[4888]: I1124 01:21:17.710545 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn9q7\" (UniqueName: \"kubernetes.io/projected/2d25d7f9-02eb-4f78-a3c9-2f9a9a533797-kube-api-access-cn9q7\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.003709 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" event={"ID":"2d25d7f9-02eb-4f78-a3c9-2f9a9a533797","Type":"ContainerDied","Data":"dffd3df3131caa1ce2d69945f16927a5f3483a1b83a868268cf5cb3ff27d5d4d"} Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.004035 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dffd3df3131caa1ce2d69945f16927a5f3483a1b83a868268cf5cb3ff27d5d4d" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.003757 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.113708 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv"] Nov 24 01:21:18 crc kubenswrapper[4888]: E1124 01:21:18.114276 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.114302 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.114601 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d25d7f9-02eb-4f78-a3c9-2f9a9a533797" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.115627 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.120149 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.120199 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.123153 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.123229 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.123283 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.124974 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.125186 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.125340 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.125565 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.126456 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.128176 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv"] Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.219747 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp2sr\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.219789 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.219900 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.219923 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.219945 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220011 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220114 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220224 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220286 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220318 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220355 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220423 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220548 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220644 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220664 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220709 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.220771 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323352 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323406 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323437 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323503 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323530 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp2sr\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323548 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323627 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323647 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323685 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323751 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323775 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323800 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323861 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323884 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323925 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.323973 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.324057 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.328532 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.328756 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.329574 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.329847 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.329867 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.329960 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.331105 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.331999 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.332336 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.333047 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.333126 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.334224 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.334627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.337264 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.337275 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.337402 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.361395 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp2sr\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.435861 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:21:18 crc kubenswrapper[4888]: I1124 01:21:18.809308 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv"] Nov 24 01:21:18 crc kubenswrapper[4888]: W1124 01:21:18.815487 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fce69b0_15d9_41d7_94f0_2263b74dff10.slice/crio-87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04 WatchSource:0}: Error finding container 87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04: Status 404 returned error can't find the container with id 87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04 Nov 24 01:21:19 crc kubenswrapper[4888]: I1124 01:21:19.021084 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" event={"ID":"7fce69b0-15d9-41d7-94f0-2263b74dff10","Type":"ContainerStarted","Data":"87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04"} Nov 24 01:21:20 crc kubenswrapper[4888]: I1124 01:21:20.040014 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" event={"ID":"7fce69b0-15d9-41d7-94f0-2263b74dff10","Type":"ContainerStarted","Data":"8a00bb2f7b3361a9a6181a1e6c870f2da1688cdc52ce6023be64c1399d3c74b2"} Nov 24 01:21:20 crc kubenswrapper[4888]: I1124 01:21:20.083588 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" podStartSLOduration=1.6132515920000001 podStartE2EDuration="2.083567786s" podCreationTimestamp="2025-11-24 01:21:18 +0000 UTC" firstStartedPulling="2025-11-24 01:21:18.81819119 +0000 UTC m=+3381.400875234" lastFinishedPulling="2025-11-24 01:21:19.288507374 +0000 UTC m=+3381.871191428" observedRunningTime="2025-11-24 01:21:20.073424771 +0000 UTC m=+3382.656108855" watchObservedRunningTime="2025-11-24 01:21:20.083567786 +0000 UTC m=+3382.666251840" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.030562 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.032908 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.044688 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.154351 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.154795 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lb4p\" (UniqueName: \"kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.155000 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.257323 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lb4p\" (UniqueName: \"kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.257415 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.257518 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.258051 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.258057 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.276596 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lb4p\" (UniqueName: \"kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p\") pod \"redhat-operators-dvnwh\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.368661 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.591434 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.592093 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.592154 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.593324 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.593400 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d" gracePeriod=600 Nov 24 01:21:23 crc kubenswrapper[4888]: I1124 01:21:23.864960 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:23 crc kubenswrapper[4888]: W1124 01:21:23.882542 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbe7904_8c8b_4c17_92ae_a233cbc1a4e6.slice/crio-29966b8b989256e35c163df91199c2ecf8ebf624b3dc01723dc6535e449000ff WatchSource:0}: Error finding container 29966b8b989256e35c163df91199c2ecf8ebf624b3dc01723dc6535e449000ff: Status 404 returned error can't find the container with id 29966b8b989256e35c163df91199c2ecf8ebf624b3dc01723dc6535e449000ff Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.103070 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d" exitCode=0 Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.103139 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d"} Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.103700 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589"} Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.103736 4888 scope.go:117] "RemoveContainer" containerID="996e8c77d1b929dbad1149d1dec6ccac0021d1a32dcbfc708489c311167971e0" Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.106202 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerStarted","Data":"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f"} Nov 24 01:21:24 crc kubenswrapper[4888]: I1124 01:21:24.106237 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerStarted","Data":"29966b8b989256e35c163df91199c2ecf8ebf624b3dc01723dc6535e449000ff"} Nov 24 01:21:25 crc kubenswrapper[4888]: I1124 01:21:25.120305 4888 generic.go:334] "Generic (PLEG): container finished" podID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerID="b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f" exitCode=0 Nov 24 01:21:25 crc kubenswrapper[4888]: I1124 01:21:25.120423 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerDied","Data":"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f"} Nov 24 01:21:25 crc kubenswrapper[4888]: I1124 01:21:25.122116 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerStarted","Data":"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0"} Nov 24 01:21:27 crc kubenswrapper[4888]: I1124 01:21:27.150080 4888 generic.go:334] "Generic (PLEG): container finished" podID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerID="cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0" exitCode=0 Nov 24 01:21:27 crc kubenswrapper[4888]: I1124 01:21:27.150164 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerDied","Data":"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0"} Nov 24 01:21:28 crc kubenswrapper[4888]: I1124 01:21:28.165994 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerStarted","Data":"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef"} Nov 24 01:21:28 crc kubenswrapper[4888]: I1124 01:21:28.194084 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvnwh" podStartSLOduration=2.626213179 podStartE2EDuration="6.194058532s" podCreationTimestamp="2025-11-24 01:21:22 +0000 UTC" firstStartedPulling="2025-11-24 01:21:24.108882663 +0000 UTC m=+3386.691566707" lastFinishedPulling="2025-11-24 01:21:27.676728006 +0000 UTC m=+3390.259412060" observedRunningTime="2025-11-24 01:21:28.18512313 +0000 UTC m=+3390.767807204" watchObservedRunningTime="2025-11-24 01:21:28.194058532 +0000 UTC m=+3390.776742606" Nov 24 01:21:33 crc kubenswrapper[4888]: I1124 01:21:33.369207 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:33 crc kubenswrapper[4888]: I1124 01:21:33.369943 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:34 crc kubenswrapper[4888]: I1124 01:21:34.427763 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvnwh" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="registry-server" probeResult="failure" output=< Nov 24 01:21:34 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 01:21:34 crc kubenswrapper[4888]: > Nov 24 01:21:43 crc kubenswrapper[4888]: I1124 01:21:43.455214 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:43 crc kubenswrapper[4888]: I1124 01:21:43.551616 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:43 crc kubenswrapper[4888]: I1124 01:21:43.708122 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:45 crc kubenswrapper[4888]: I1124 01:21:45.362199 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dvnwh" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="registry-server" containerID="cri-o://08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef" gracePeriod=2 Nov 24 01:21:45 crc kubenswrapper[4888]: E1124 01:21:45.576343 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbe7904_8c8b_4c17_92ae_a233cbc1a4e6.slice/crio-08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbe7904_8c8b_4c17_92ae_a233cbc1a4e6.slice/crio-conmon-08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef.scope\": RecentStats: unable to find data in memory cache]" Nov 24 01:21:45 crc kubenswrapper[4888]: I1124 01:21:45.947243 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.113284 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lb4p\" (UniqueName: \"kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p\") pod \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.113436 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content\") pod \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.113506 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities\") pod \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\" (UID: \"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6\") " Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.114186 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities" (OuterVolumeSpecName: "utilities") pod "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" (UID: "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.119068 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p" (OuterVolumeSpecName: "kube-api-access-5lb4p") pod "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" (UID: "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6"). InnerVolumeSpecName "kube-api-access-5lb4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.215970 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.216000 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lb4p\" (UniqueName: \"kubernetes.io/projected/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-kube-api-access-5lb4p\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.217102 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" (UID: "fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.319031 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.377697 4888 generic.go:334] "Generic (PLEG): container finished" podID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerID="08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef" exitCode=0 Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.377755 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvnwh" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.377765 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerDied","Data":"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef"} Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.377802 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvnwh" event={"ID":"fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6","Type":"ContainerDied","Data":"29966b8b989256e35c163df91199c2ecf8ebf624b3dc01723dc6535e449000ff"} Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.377839 4888 scope.go:117] "RemoveContainer" containerID="08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.408838 4888 scope.go:117] "RemoveContainer" containerID="cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.414601 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.426560 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dvnwh"] Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.433083 4888 scope.go:117] "RemoveContainer" containerID="b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.494329 4888 scope.go:117] "RemoveContainer" containerID="08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef" Nov 24 01:21:46 crc kubenswrapper[4888]: E1124 01:21:46.494779 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef\": container with ID starting with 08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef not found: ID does not exist" containerID="08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.494864 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef"} err="failed to get container status \"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef\": rpc error: code = NotFound desc = could not find container \"08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef\": container with ID starting with 08d8dbb73a848646c1c8119dbae41e560662077ecc51903f33a84824259218ef not found: ID does not exist" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.494904 4888 scope.go:117] "RemoveContainer" containerID="cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0" Nov 24 01:21:46 crc kubenswrapper[4888]: E1124 01:21:46.495376 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0\": container with ID starting with cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0 not found: ID does not exist" containerID="cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.495410 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0"} err="failed to get container status \"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0\": rpc error: code = NotFound desc = could not find container \"cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0\": container with ID starting with cc4a016d86909da9c2dd6752a39ea9061ba35e39c3ff03160189cdfce756bfc0 not found: ID does not exist" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.495440 4888 scope.go:117] "RemoveContainer" containerID="b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f" Nov 24 01:21:46 crc kubenswrapper[4888]: E1124 01:21:46.495765 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f\": container with ID starting with b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f not found: ID does not exist" containerID="b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f" Nov 24 01:21:46 crc kubenswrapper[4888]: I1124 01:21:46.495887 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f"} err="failed to get container status \"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f\": rpc error: code = NotFound desc = could not find container \"b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f\": container with ID starting with b875aefd039e0c38e17ef26a3e95d547141473d1bd264b97b3bff8353484580f not found: ID does not exist" Nov 24 01:21:48 crc kubenswrapper[4888]: I1124 01:21:48.262587 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" path="/var/lib/kubelet/pods/fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6/volumes" Nov 24 01:22:20 crc kubenswrapper[4888]: I1124 01:22:20.842673 4888 generic.go:334] "Generic (PLEG): container finished" podID="7fce69b0-15d9-41d7-94f0-2263b74dff10" containerID="8a00bb2f7b3361a9a6181a1e6c870f2da1688cdc52ce6023be64c1399d3c74b2" exitCode=0 Nov 24 01:22:20 crc kubenswrapper[4888]: I1124 01:22:20.842798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" event={"ID":"7fce69b0-15d9-41d7-94f0-2263b74dff10","Type":"ContainerDied","Data":"8a00bb2f7b3361a9a6181a1e6c870f2da1688cdc52ce6023be64c1399d3c74b2"} Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.346897 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475112 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475426 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475513 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475621 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475732 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475840 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.475975 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.476070 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp2sr\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.476191 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.476279 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.476401 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.476545 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.477064 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.477591 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.477729 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.477840 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.477949 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle\") pod \"7fce69b0-15d9-41d7-94f0-2263b74dff10\" (UID: \"7fce69b0-15d9-41d7-94f0-2263b74dff10\") " Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.483996 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.484284 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.484751 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph" (OuterVolumeSpecName: "ceph") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.484827 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.485256 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.485375 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.486168 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.487523 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.488634 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.489129 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.490056 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.491487 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.492181 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr" (OuterVolumeSpecName: "kube-api-access-hp2sr") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "kube-api-access-hp2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.492885 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.493529 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.522752 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory" (OuterVolumeSpecName: "inventory") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.523213 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fce69b0-15d9-41d7-94f0-2263b74dff10" (UID: "7fce69b0-15d9-41d7-94f0-2263b74dff10"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581635 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581669 4888 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581683 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581694 4888 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581705 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581714 4888 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581723 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581735 4888 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581743 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581752 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581760 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581770 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581779 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581788 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581796 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581805 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce69b0-15d9-41d7-94f0-2263b74dff10-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.581835 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp2sr\" (UniqueName: \"kubernetes.io/projected/7fce69b0-15d9-41d7-94f0-2263b74dff10-kube-api-access-hp2sr\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.870318 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" event={"ID":"7fce69b0-15d9-41d7-94f0-2263b74dff10","Type":"ContainerDied","Data":"87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04"} Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.870564 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87b7fc519ef29dfd9b34d903cc79f3516bc59ef5dcf64447ca47bd22db3e8a04" Nov 24 01:22:22 crc kubenswrapper[4888]: I1124 01:22:22.870525 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.001466 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv"] Nov 24 01:22:23 crc kubenswrapper[4888]: E1124 01:22:23.002452 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="extract-utilities" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.002619 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="extract-utilities" Nov 24 01:22:23 crc kubenswrapper[4888]: E1124 01:22:23.002760 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="extract-content" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.002902 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="extract-content" Nov 24 01:22:23 crc kubenswrapper[4888]: E1124 01:22:23.003061 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fce69b0-15d9-41d7-94f0-2263b74dff10" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.003181 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fce69b0-15d9-41d7-94f0-2263b74dff10" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:23 crc kubenswrapper[4888]: E1124 01:22:23.003324 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="registry-server" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.003432 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="registry-server" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.003781 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fce69b0-15d9-41d7-94f0-2263b74dff10" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.003854 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbbe7904-8c8b-4c17-92ae-a233cbc1a4e6" containerName="registry-server" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.004687 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.007647 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.008142 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.008606 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.007651 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.012047 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv"] Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.013598 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.091870 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.092159 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbkt\" (UniqueName: \"kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.092297 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.092434 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.195158 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.195541 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbkt\" (UniqueName: \"kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.195726 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.196414 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.199910 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.200416 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.201594 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.218871 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbkt\" (UniqueName: \"kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.327383 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:23 crc kubenswrapper[4888]: I1124 01:22:23.999094 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv"] Nov 24 01:22:24 crc kubenswrapper[4888]: W1124 01:22:24.001835 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea89083c_a0ce_4bfe_b419_be89f9643924.slice/crio-48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef WatchSource:0}: Error finding container 48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef: Status 404 returned error can't find the container with id 48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef Nov 24 01:22:24 crc kubenswrapper[4888]: I1124 01:22:24.004885 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:22:24 crc kubenswrapper[4888]: I1124 01:22:24.894664 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" event={"ID":"ea89083c-a0ce-4bfe-b419-be89f9643924","Type":"ContainerStarted","Data":"b5a4d317257c1c86e4d90fc859d0c69877b3b3fcc4429568270d145d8b4511af"} Nov 24 01:22:24 crc kubenswrapper[4888]: I1124 01:22:24.894997 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" event={"ID":"ea89083c-a0ce-4bfe-b419-be89f9643924","Type":"ContainerStarted","Data":"48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef"} Nov 24 01:22:24 crc kubenswrapper[4888]: I1124 01:22:24.920005 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" podStartSLOduration=2.365024689 podStartE2EDuration="2.919981535s" podCreationTimestamp="2025-11-24 01:22:22 +0000 UTC" firstStartedPulling="2025-11-24 01:22:24.004658239 +0000 UTC m=+3446.587342283" lastFinishedPulling="2025-11-24 01:22:24.559615085 +0000 UTC m=+3447.142299129" observedRunningTime="2025-11-24 01:22:24.914133951 +0000 UTC m=+3447.496817985" watchObservedRunningTime="2025-11-24 01:22:24.919981535 +0000 UTC m=+3447.502665579" Nov 24 01:22:31 crc kubenswrapper[4888]: I1124 01:22:31.972875 4888 generic.go:334] "Generic (PLEG): container finished" podID="ea89083c-a0ce-4bfe-b419-be89f9643924" containerID="b5a4d317257c1c86e4d90fc859d0c69877b3b3fcc4429568270d145d8b4511af" exitCode=0 Nov 24 01:22:31 crc kubenswrapper[4888]: I1124 01:22:31.972991 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" event={"ID":"ea89083c-a0ce-4bfe-b419-be89f9643924","Type":"ContainerDied","Data":"b5a4d317257c1c86e4d90fc859d0c69877b3b3fcc4429568270d145d8b4511af"} Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.471856 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.538880 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key\") pod \"ea89083c-a0ce-4bfe-b419-be89f9643924\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.538992 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory\") pod \"ea89083c-a0ce-4bfe-b419-be89f9643924\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.539109 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph\") pod \"ea89083c-a0ce-4bfe-b419-be89f9643924\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.539186 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxbkt\" (UniqueName: \"kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt\") pod \"ea89083c-a0ce-4bfe-b419-be89f9643924\" (UID: \"ea89083c-a0ce-4bfe-b419-be89f9643924\") " Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.545581 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt" (OuterVolumeSpecName: "kube-api-access-gxbkt") pod "ea89083c-a0ce-4bfe-b419-be89f9643924" (UID: "ea89083c-a0ce-4bfe-b419-be89f9643924"). InnerVolumeSpecName "kube-api-access-gxbkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.546188 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph" (OuterVolumeSpecName: "ceph") pod "ea89083c-a0ce-4bfe-b419-be89f9643924" (UID: "ea89083c-a0ce-4bfe-b419-be89f9643924"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.592653 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory" (OuterVolumeSpecName: "inventory") pod "ea89083c-a0ce-4bfe-b419-be89f9643924" (UID: "ea89083c-a0ce-4bfe-b419-be89f9643924"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.593348 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea89083c-a0ce-4bfe-b419-be89f9643924" (UID: "ea89083c-a0ce-4bfe-b419-be89f9643924"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.641234 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxbkt\" (UniqueName: \"kubernetes.io/projected/ea89083c-a0ce-4bfe-b419-be89f9643924-kube-api-access-gxbkt\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.641267 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.641278 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.641287 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea89083c-a0ce-4bfe-b419-be89f9643924-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.995951 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" event={"ID":"ea89083c-a0ce-4bfe-b419-be89f9643924","Type":"ContainerDied","Data":"48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef"} Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.996383 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48e00563e7017f91b2aaf8d74391338c175a966c8b09141673a84299287216ef" Nov 24 01:22:33 crc kubenswrapper[4888]: I1124 01:22:33.996013 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.198610 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp"] Nov 24 01:22:34 crc kubenswrapper[4888]: E1124 01:22:34.199790 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea89083c-a0ce-4bfe-b419-be89f9643924" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.199844 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea89083c-a0ce-4bfe-b419-be89f9643924" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.200282 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea89083c-a0ce-4bfe-b419-be89f9643924" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.201708 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.204429 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.204660 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.204879 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.205344 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.205390 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.205552 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.213739 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp"] Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.259644 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.259742 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.259852 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.259911 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.259974 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.260029 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmktm\" (UniqueName: \"kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361679 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361752 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361835 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361887 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361949 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.361984 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmktm\" (UniqueName: \"kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.362979 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.367025 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.367201 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.367493 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.367893 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.382609 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmktm\" (UniqueName: \"kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gjwnp\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:34 crc kubenswrapper[4888]: I1124 01:22:34.534725 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:22:35 crc kubenswrapper[4888]: I1124 01:22:35.131164 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp"] Nov 24 01:22:36 crc kubenswrapper[4888]: I1124 01:22:36.024049 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" event={"ID":"26532baf-63c3-4d2f-87c2-9c690c53823c","Type":"ContainerStarted","Data":"348542a7586aaa7dad9bda10abbb8e6afc064c9f37f8c6e3787bcdf734458941"} Nov 24 01:22:36 crc kubenswrapper[4888]: I1124 01:22:36.024509 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" event={"ID":"26532baf-63c3-4d2f-87c2-9c690c53823c","Type":"ContainerStarted","Data":"a25d2c8af42b732f8029b014904385ec05aa9aba081a6c66e5d6697ecffabd57"} Nov 24 01:22:36 crc kubenswrapper[4888]: I1124 01:22:36.066562 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" podStartSLOduration=1.645644406 podStartE2EDuration="2.06653561s" podCreationTimestamp="2025-11-24 01:22:34 +0000 UTC" firstStartedPulling="2025-11-24 01:22:35.129335329 +0000 UTC m=+3457.712019383" lastFinishedPulling="2025-11-24 01:22:35.550226503 +0000 UTC m=+3458.132910587" observedRunningTime="2025-11-24 01:22:36.050360515 +0000 UTC m=+3458.633044599" watchObservedRunningTime="2025-11-24 01:22:36.06653561 +0000 UTC m=+3458.649219664" Nov 24 01:23:14 crc kubenswrapper[4888]: I1124 01:23:14.347166 4888 scope.go:117] "RemoveContainer" containerID="bac6aac4dacbd78b69913fa4fcf2433d905a370d9406f2b306700347e3c907cf" Nov 24 01:23:14 crc kubenswrapper[4888]: I1124 01:23:14.395840 4888 scope.go:117] "RemoveContainer" containerID="4c5bf4d62a53585e9ccb5f751c8c202752965c9809c3409595e0c33f78f07118" Nov 24 01:23:14 crc kubenswrapper[4888]: I1124 01:23:14.450197 4888 scope.go:117] "RemoveContainer" containerID="a6bffa3456299327e90be492361c2c14c78a33ba026b7f15ee91575dc0a8b3de" Nov 24 01:23:23 crc kubenswrapper[4888]: I1124 01:23:23.591563 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:23:23 crc kubenswrapper[4888]: I1124 01:23:23.592252 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:23:52 crc kubenswrapper[4888]: I1124 01:23:52.954001 4888 generic.go:334] "Generic (PLEG): container finished" podID="26532baf-63c3-4d2f-87c2-9c690c53823c" containerID="348542a7586aaa7dad9bda10abbb8e6afc064c9f37f8c6e3787bcdf734458941" exitCode=0 Nov 24 01:23:52 crc kubenswrapper[4888]: I1124 01:23:52.954085 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" event={"ID":"26532baf-63c3-4d2f-87c2-9c690c53823c","Type":"ContainerDied","Data":"348542a7586aaa7dad9bda10abbb8e6afc064c9f37f8c6e3787bcdf734458941"} Nov 24 01:23:53 crc kubenswrapper[4888]: I1124 01:23:53.592429 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:23:53 crc kubenswrapper[4888]: I1124 01:23:53.592499 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.545969 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636484 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636728 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636765 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmktm\" (UniqueName: \"kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636790 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636860 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.636937 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0\") pod \"26532baf-63c3-4d2f-87c2-9c690c53823c\" (UID: \"26532baf-63c3-4d2f-87c2-9c690c53823c\") " Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.642320 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph" (OuterVolumeSpecName: "ceph") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.644181 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm" (OuterVolumeSpecName: "kube-api-access-cmktm") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "kube-api-access-cmktm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.644976 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.675476 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.677109 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory" (OuterVolumeSpecName: "inventory") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.690988 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "26532baf-63c3-4d2f-87c2-9c690c53823c" (UID: "26532baf-63c3-4d2f-87c2-9c690c53823c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740695 4888 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/26532baf-63c3-4d2f-87c2-9c690c53823c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740762 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740788 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740843 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmktm\" (UniqueName: \"kubernetes.io/projected/26532baf-63c3-4d2f-87c2-9c690c53823c-kube-api-access-cmktm\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740867 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.740891 4888 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26532baf-63c3-4d2f-87c2-9c690c53823c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.988448 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" event={"ID":"26532baf-63c3-4d2f-87c2-9c690c53823c","Type":"ContainerDied","Data":"a25d2c8af42b732f8029b014904385ec05aa9aba081a6c66e5d6697ecffabd57"} Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.988491 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25d2c8af42b732f8029b014904385ec05aa9aba081a6c66e5d6697ecffabd57" Nov 24 01:23:54 crc kubenswrapper[4888]: I1124 01:23:54.988544 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gjwnp" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.101338 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv"] Nov 24 01:23:55 crc kubenswrapper[4888]: E1124 01:23:55.101892 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26532baf-63c3-4d2f-87c2-9c690c53823c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.101910 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="26532baf-63c3-4d2f-87c2-9c690c53823c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.102195 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="26532baf-63c3-4d2f-87c2-9c690c53823c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.103156 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.108796 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109090 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109239 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109380 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109521 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109666 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.109692 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.122778 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv"] Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148274 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148336 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148366 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148419 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6s9g\" (UniqueName: \"kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148481 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.148594 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250076 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250293 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250337 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250364 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250396 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250427 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6s9g\" (UniqueName: \"kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.250454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.254188 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.257217 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.258995 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.260188 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.260943 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.261233 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.274880 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6s9g\" (UniqueName: \"kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:55 crc kubenswrapper[4888]: I1124 01:23:55.440796 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:23:56 crc kubenswrapper[4888]: I1124 01:23:56.087693 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv"] Nov 24 01:23:57 crc kubenswrapper[4888]: I1124 01:23:57.019411 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" event={"ID":"fb3242a5-393f-4fda-a791-0d1246a5bded","Type":"ContainerStarted","Data":"e638878bf1acb99d9675a5d11f668840de2e5110cdd21a4750c7da62be1160c1"} Nov 24 01:23:57 crc kubenswrapper[4888]: I1124 01:23:57.022032 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" event={"ID":"fb3242a5-393f-4fda-a791-0d1246a5bded","Type":"ContainerStarted","Data":"f08a0ac3413115768807e1bf975c68c5d6281830986b5b154e62731ea5d5d8e1"} Nov 24 01:23:57 crc kubenswrapper[4888]: I1124 01:23:57.058047 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" podStartSLOduration=1.627736418 podStartE2EDuration="2.058015115s" podCreationTimestamp="2025-11-24 01:23:55 +0000 UTC" firstStartedPulling="2025-11-24 01:23:56.094659368 +0000 UTC m=+3538.677343452" lastFinishedPulling="2025-11-24 01:23:56.524938105 +0000 UTC m=+3539.107622149" observedRunningTime="2025-11-24 01:23:57.046148741 +0000 UTC m=+3539.628832805" watchObservedRunningTime="2025-11-24 01:23:57.058015115 +0000 UTC m=+3539.640699199" Nov 24 01:24:23 crc kubenswrapper[4888]: I1124 01:24:23.591649 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:24:23 crc kubenswrapper[4888]: I1124 01:24:23.592347 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:24:23 crc kubenswrapper[4888]: I1124 01:24:23.592402 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:24:23 crc kubenswrapper[4888]: I1124 01:24:23.593273 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:24:23 crc kubenswrapper[4888]: I1124 01:24:23.593341 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" gracePeriod=600 Nov 24 01:24:23 crc kubenswrapper[4888]: E1124 01:24:23.750461 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:24:24 crc kubenswrapper[4888]: I1124 01:24:24.334402 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" exitCode=0 Nov 24 01:24:24 crc kubenswrapper[4888]: I1124 01:24:24.334508 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589"} Nov 24 01:24:24 crc kubenswrapper[4888]: I1124 01:24:24.335143 4888 scope.go:117] "RemoveContainer" containerID="e1982dd59027dcb6a64099f9a16f62e68b2cc60becc33cb22b794b3c1010978d" Nov 24 01:24:24 crc kubenswrapper[4888]: I1124 01:24:24.336221 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:24:24 crc kubenswrapper[4888]: E1124 01:24:24.336777 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:24:35 crc kubenswrapper[4888]: I1124 01:24:35.245426 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:24:35 crc kubenswrapper[4888]: E1124 01:24:35.246474 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:24:48 crc kubenswrapper[4888]: I1124 01:24:48.257230 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:24:48 crc kubenswrapper[4888]: E1124 01:24:48.258271 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:25:02 crc kubenswrapper[4888]: I1124 01:25:02.245788 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:25:02 crc kubenswrapper[4888]: E1124 01:25:02.247916 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:25:12 crc kubenswrapper[4888]: I1124 01:25:12.889223 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb3242a5-393f-4fda-a791-0d1246a5bded" containerID="e638878bf1acb99d9675a5d11f668840de2e5110cdd21a4750c7da62be1160c1" exitCode=0 Nov 24 01:25:12 crc kubenswrapper[4888]: I1124 01:25:12.889349 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" event={"ID":"fb3242a5-393f-4fda-a791-0d1246a5bded","Type":"ContainerDied","Data":"e638878bf1acb99d9675a5d11f668840de2e5110cdd21a4750c7da62be1160c1"} Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.416869 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506265 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506314 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506344 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506385 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506527 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506551 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6s9g\" (UniqueName: \"kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.506599 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0\") pod \"fb3242a5-393f-4fda-a791-0d1246a5bded\" (UID: \"fb3242a5-393f-4fda-a791-0d1246a5bded\") " Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.513616 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.515229 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph" (OuterVolumeSpecName: "ceph") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.521647 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g" (OuterVolumeSpecName: "kube-api-access-h6s9g") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "kube-api-access-h6s9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.545200 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.547750 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory" (OuterVolumeSpecName: "inventory") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.553205 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.562755 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb3242a5-393f-4fda-a791-0d1246a5bded" (UID: "fb3242a5-393f-4fda-a791-0d1246a5bded"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.567196 4888 scope.go:117] "RemoveContainer" containerID="dd3261c101da221faf72a8c9044abe4af380ab78c079fca39c0a17f8025a280b" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.611116 4888 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.611474 4888 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.611633 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6s9g\" (UniqueName: \"kubernetes.io/projected/fb3242a5-393f-4fda-a791-0d1246a5bded-kube-api-access-h6s9g\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.611755 4888 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.611922 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.612142 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.612321 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb3242a5-393f-4fda-a791-0d1246a5bded-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.642615 4888 scope.go:117] "RemoveContainer" containerID="ec95381f28ac1a8963145df04186747df58234f2cab9256cfd1e50c09e10a4fd" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.661866 4888 scope.go:117] "RemoveContainer" containerID="435e15b253f9bd4dd87501abfa2d82abe5fbec82234beae2d49865ba59acc775" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.919220 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" event={"ID":"fb3242a5-393f-4fda-a791-0d1246a5bded","Type":"ContainerDied","Data":"f08a0ac3413115768807e1bf975c68c5d6281830986b5b154e62731ea5d5d8e1"} Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.919588 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f08a0ac3413115768807e1bf975c68c5d6281830986b5b154e62731ea5d5d8e1" Nov 24 01:25:14 crc kubenswrapper[4888]: I1124 01:25:14.919371 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.223550 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4"] Nov 24 01:25:15 crc kubenswrapper[4888]: E1124 01:25:15.224211 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3242a5-393f-4fda-a791-0d1246a5bded" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.224236 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3242a5-393f-4fda-a791-0d1246a5bded" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.224539 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3242a5-393f-4fda-a791-0d1246a5bded" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.225875 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.233898 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.234019 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.234110 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.234179 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.234477 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.241499 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4"] Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.242835 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.327555 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.327642 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.327718 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.328149 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.328388 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9l2p\" (UniqueName: \"kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.328455 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430521 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9l2p\" (UniqueName: \"kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430589 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430732 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430785 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.430973 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.436880 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.437011 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.437382 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.437435 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.440587 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.448658 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9l2p\" (UniqueName: \"kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:15 crc kubenswrapper[4888]: I1124 01:25:15.565519 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:25:16 crc kubenswrapper[4888]: I1124 01:25:16.111781 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4"] Nov 24 01:25:16 crc kubenswrapper[4888]: W1124 01:25:16.113981 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4e89186_7454_4e0c_85ed_f7c90f6a7b0f.slice/crio-9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be WatchSource:0}: Error finding container 9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be: Status 404 returned error can't find the container with id 9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be Nov 24 01:25:16 crc kubenswrapper[4888]: I1124 01:25:16.945382 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" event={"ID":"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f","Type":"ContainerStarted","Data":"b52e9b74e9408aa6744eedac58e351527bfd1074c0c608eb34f8985859d047b9"} Nov 24 01:25:16 crc kubenswrapper[4888]: I1124 01:25:16.946233 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" event={"ID":"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f","Type":"ContainerStarted","Data":"9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be"} Nov 24 01:25:16 crc kubenswrapper[4888]: I1124 01:25:16.971529 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" podStartSLOduration=1.485770327 podStartE2EDuration="1.971497111s" podCreationTimestamp="2025-11-24 01:25:15 +0000 UTC" firstStartedPulling="2025-11-24 01:25:16.117033438 +0000 UTC m=+3618.699717482" lastFinishedPulling="2025-11-24 01:25:16.602760222 +0000 UTC m=+3619.185444266" observedRunningTime="2025-11-24 01:25:16.963564448 +0000 UTC m=+3619.546248502" watchObservedRunningTime="2025-11-24 01:25:16.971497111 +0000 UTC m=+3619.554181175" Nov 24 01:25:17 crc kubenswrapper[4888]: I1124 01:25:17.245837 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:25:17 crc kubenswrapper[4888]: E1124 01:25:17.246153 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:25:28 crc kubenswrapper[4888]: I1124 01:25:28.251807 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:25:28 crc kubenswrapper[4888]: E1124 01:25:28.252905 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:25:43 crc kubenswrapper[4888]: I1124 01:25:43.245328 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:25:43 crc kubenswrapper[4888]: E1124 01:25:43.246149 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:25:58 crc kubenswrapper[4888]: I1124 01:25:58.254460 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:25:58 crc kubenswrapper[4888]: E1124 01:25:58.256233 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:26:12 crc kubenswrapper[4888]: I1124 01:26:12.246029 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:26:12 crc kubenswrapper[4888]: E1124 01:26:12.247265 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:26:25 crc kubenswrapper[4888]: I1124 01:26:25.245072 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:26:25 crc kubenswrapper[4888]: E1124 01:26:25.246919 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:26:40 crc kubenswrapper[4888]: I1124 01:26:40.245043 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:26:40 crc kubenswrapper[4888]: E1124 01:26:40.245954 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:26:54 crc kubenswrapper[4888]: I1124 01:26:54.246186 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:26:54 crc kubenswrapper[4888]: E1124 01:26:54.247155 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:27:07 crc kubenswrapper[4888]: I1124 01:27:07.245647 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:27:07 crc kubenswrapper[4888]: E1124 01:27:07.246530 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:27:20 crc kubenswrapper[4888]: I1124 01:27:20.246416 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:27:20 crc kubenswrapper[4888]: E1124 01:27:20.247741 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:27:31 crc kubenswrapper[4888]: I1124 01:27:31.245602 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:27:31 crc kubenswrapper[4888]: E1124 01:27:31.246486 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:27:42 crc kubenswrapper[4888]: I1124 01:27:42.246638 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:27:42 crc kubenswrapper[4888]: E1124 01:27:42.247736 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:27:54 crc kubenswrapper[4888]: I1124 01:27:54.245641 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:27:54 crc kubenswrapper[4888]: E1124 01:27:54.246771 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:28:07 crc kubenswrapper[4888]: I1124 01:28:07.245982 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:28:07 crc kubenswrapper[4888]: E1124 01:28:07.246774 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:28:22 crc kubenswrapper[4888]: I1124 01:28:22.121168 4888 generic.go:334] "Generic (PLEG): container finished" podID="c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" containerID="b52e9b74e9408aa6744eedac58e351527bfd1074c0c608eb34f8985859d047b9" exitCode=0 Nov 24 01:28:22 crc kubenswrapper[4888]: I1124 01:28:22.121263 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" event={"ID":"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f","Type":"ContainerDied","Data":"b52e9b74e9408aa6744eedac58e351527bfd1074c0c608eb34f8985859d047b9"} Nov 24 01:28:22 crc kubenswrapper[4888]: I1124 01:28:22.247346 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:28:22 crc kubenswrapper[4888]: E1124 01:28:22.247708 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.725554 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.838562 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9l2p\" (UniqueName: \"kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.838716 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.838935 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.838975 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.839008 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.839065 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle\") pod \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\" (UID: \"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f\") " Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.845402 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p" (OuterVolumeSpecName: "kube-api-access-z9l2p") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "kube-api-access-z9l2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.847032 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph" (OuterVolumeSpecName: "ceph") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.850079 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.880176 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.884666 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.885416 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory" (OuterVolumeSpecName: "inventory") pod "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" (UID: "c4e89186-7454-4e0c-85ed-f7c90f6a7b0f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941193 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941232 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941246 4888 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941258 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9l2p\" (UniqueName: \"kubernetes.io/projected/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-kube-api-access-z9l2p\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941271 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:23 crc kubenswrapper[4888]: I1124 01:28:23.941281 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4e89186-7454-4e0c-85ed-f7c90f6a7b0f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.148706 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" event={"ID":"c4e89186-7454-4e0c-85ed-f7c90f6a7b0f","Type":"ContainerDied","Data":"9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be"} Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.149129 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9877d31d980dcdf6a485df55e3e7a379e94583ffa2e0b54b61defa07f91726be" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.148908 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.282989 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt"] Nov 24 01:28:24 crc kubenswrapper[4888]: E1124 01:28:24.283589 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.283618 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.284006 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e89186-7454-4e0c-85ed-f7c90f6a7b0f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.285185 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.287917 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.287922 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.288472 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.288666 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.288715 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.288853 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.295766 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.296040 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.296610 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt"] Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.296788 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451170 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451255 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451480 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451570 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5w6j\" (UniqueName: \"kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451909 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.451999 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.452036 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.452152 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.452273 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.452315 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.452398 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.554822 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.554901 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.554964 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.554991 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5w6j\" (UniqueName: \"kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555082 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555119 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555145 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555186 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555228 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555254 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.555288 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.556340 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.557384 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.559922 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.560217 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.560758 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.559581 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.561458 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.561902 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.562543 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.562894 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.574217 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5w6j\" (UniqueName: \"kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:24 crc kubenswrapper[4888]: I1124 01:28:24.612699 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:28:25 crc kubenswrapper[4888]: I1124 01:28:25.248317 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt"] Nov 24 01:28:25 crc kubenswrapper[4888]: I1124 01:28:25.251759 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:28:26 crc kubenswrapper[4888]: I1124 01:28:26.174093 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" event={"ID":"59680387-8e94-4e51-9b44-9c0829e262a6","Type":"ContainerStarted","Data":"feef08808100513068d122b513e7db053d166688a6790804811c018fe692d31e"} Nov 24 01:28:26 crc kubenswrapper[4888]: I1124 01:28:26.174704 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" event={"ID":"59680387-8e94-4e51-9b44-9c0829e262a6","Type":"ContainerStarted","Data":"35be167aabfb6cfdefe750243ef233b4faca0f54c9b9536ea986ddd0cd9d67cc"} Nov 24 01:28:26 crc kubenswrapper[4888]: I1124 01:28:26.204349 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" podStartSLOduration=1.639078312 podStartE2EDuration="2.2043286s" podCreationTimestamp="2025-11-24 01:28:24 +0000 UTC" firstStartedPulling="2025-11-24 01:28:25.251428323 +0000 UTC m=+3807.834112377" lastFinishedPulling="2025-11-24 01:28:25.816678621 +0000 UTC m=+3808.399362665" observedRunningTime="2025-11-24 01:28:26.194306429 +0000 UTC m=+3808.776990493" watchObservedRunningTime="2025-11-24 01:28:26.2043286 +0000 UTC m=+3808.787012654" Nov 24 01:28:33 crc kubenswrapper[4888]: I1124 01:28:33.245491 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:28:33 crc kubenswrapper[4888]: E1124 01:28:33.246560 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:28:45 crc kubenswrapper[4888]: I1124 01:28:45.245911 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:28:45 crc kubenswrapper[4888]: E1124 01:28:45.246758 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:29:00 crc kubenswrapper[4888]: I1124 01:29:00.245890 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:29:00 crc kubenswrapper[4888]: E1124 01:29:00.246937 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:29:14 crc kubenswrapper[4888]: I1124 01:29:14.246316 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:29:14 crc kubenswrapper[4888]: E1124 01:29:14.247550 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:29:25 crc kubenswrapper[4888]: I1124 01:29:25.245446 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:29:25 crc kubenswrapper[4888]: I1124 01:29:25.871983 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf"} Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.153548 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj"] Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.156538 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.160768 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.160871 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.169569 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj"] Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.253019 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlngk\" (UniqueName: \"kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.253274 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.253402 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.355582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlngk\" (UniqueName: \"kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.355753 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.355847 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.356833 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.363337 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.379980 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlngk\" (UniqueName: \"kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk\") pod \"collect-profiles-29399130-g9hsj\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.480267 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:00 crc kubenswrapper[4888]: I1124 01:30:00.964735 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj"] Nov 24 01:30:01 crc kubenswrapper[4888]: I1124 01:30:01.328641 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" event={"ID":"3d9ece5e-18a0-4880-ad86-1764daf3584a","Type":"ContainerStarted","Data":"da4d8a886c63611947a0e9119a3c94a4c7aee5d3c20e4b8b475fb834ec9da1d0"} Nov 24 01:30:01 crc kubenswrapper[4888]: I1124 01:30:01.329017 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" event={"ID":"3d9ece5e-18a0-4880-ad86-1764daf3584a","Type":"ContainerStarted","Data":"8d8e018736c0f84f087806b958486f089338213b6b981b13c94b6bab2f118c06"} Nov 24 01:30:01 crc kubenswrapper[4888]: I1124 01:30:01.347649 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" podStartSLOduration=1.347627354 podStartE2EDuration="1.347627354s" podCreationTimestamp="2025-11-24 01:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:30:01.347172161 +0000 UTC m=+3903.929856245" watchObservedRunningTime="2025-11-24 01:30:01.347627354 +0000 UTC m=+3903.930311398" Nov 24 01:30:02 crc kubenswrapper[4888]: I1124 01:30:02.347460 4888 generic.go:334] "Generic (PLEG): container finished" podID="3d9ece5e-18a0-4880-ad86-1764daf3584a" containerID="da4d8a886c63611947a0e9119a3c94a4c7aee5d3c20e4b8b475fb834ec9da1d0" exitCode=0 Nov 24 01:30:02 crc kubenswrapper[4888]: I1124 01:30:02.347550 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" event={"ID":"3d9ece5e-18a0-4880-ad86-1764daf3584a","Type":"ContainerDied","Data":"da4d8a886c63611947a0e9119a3c94a4c7aee5d3c20e4b8b475fb834ec9da1d0"} Nov 24 01:30:03 crc kubenswrapper[4888]: I1124 01:30:03.851754 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.044585 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlngk\" (UniqueName: \"kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk\") pod \"3d9ece5e-18a0-4880-ad86-1764daf3584a\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.044682 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume\") pod \"3d9ece5e-18a0-4880-ad86-1764daf3584a\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.044888 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume\") pod \"3d9ece5e-18a0-4880-ad86-1764daf3584a\" (UID: \"3d9ece5e-18a0-4880-ad86-1764daf3584a\") " Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.045982 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d9ece5e-18a0-4880-ad86-1764daf3584a" (UID: "3d9ece5e-18a0-4880-ad86-1764daf3584a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.053348 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk" (OuterVolumeSpecName: "kube-api-access-dlngk") pod "3d9ece5e-18a0-4880-ad86-1764daf3584a" (UID: "3d9ece5e-18a0-4880-ad86-1764daf3584a"). InnerVolumeSpecName "kube-api-access-dlngk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.053506 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d9ece5e-18a0-4880-ad86-1764daf3584a" (UID: "3d9ece5e-18a0-4880-ad86-1764daf3584a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.148441 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d9ece5e-18a0-4880-ad86-1764daf3584a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.148490 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlngk\" (UniqueName: \"kubernetes.io/projected/3d9ece5e-18a0-4880-ad86-1764daf3584a-kube-api-access-dlngk\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.148506 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d9ece5e-18a0-4880-ad86-1764daf3584a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.377460 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.377264 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj" event={"ID":"3d9ece5e-18a0-4880-ad86-1764daf3584a","Type":"ContainerDied","Data":"8d8e018736c0f84f087806b958486f089338213b6b981b13c94b6bab2f118c06"} Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.377588 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d8e018736c0f84f087806b958486f089338213b6b981b13c94b6bab2f118c06" Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.436558 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47"] Nov 24 01:30:04 crc kubenswrapper[4888]: I1124 01:30:04.444985 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399085-wzc47"] Nov 24 01:30:06 crc kubenswrapper[4888]: I1124 01:30:06.267316 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8584e6b2-f5e9-4395-a596-33ead8945cf0" path="/var/lib/kubelet/pods/8584e6b2-f5e9-4395-a596-33ead8945cf0/volumes" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.790545 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:07 crc kubenswrapper[4888]: E1124 01:30:07.791807 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9ece5e-18a0-4880-ad86-1764daf3584a" containerName="collect-profiles" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.791844 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9ece5e-18a0-4880-ad86-1764daf3584a" containerName="collect-profiles" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.792160 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9ece5e-18a0-4880-ad86-1764daf3584a" containerName="collect-profiles" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.795916 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.803551 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.940697 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.941181 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:07 crc kubenswrapper[4888]: I1124 01:30:07.941781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzdrp\" (UniqueName: \"kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.044724 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzdrp\" (UniqueName: \"kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.044852 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.044959 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.045573 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.045674 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.069788 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzdrp\" (UniqueName: \"kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp\") pod \"certified-operators-vdrbf\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.126947 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:08 crc kubenswrapper[4888]: I1124 01:30:08.648107 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:09 crc kubenswrapper[4888]: I1124 01:30:09.441064 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerDied","Data":"9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59"} Nov 24 01:30:09 crc kubenswrapper[4888]: I1124 01:30:09.440804 4888 generic.go:334] "Generic (PLEG): container finished" podID="3fb76208-02ad-443b-a757-7deb1d079541" containerID="9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59" exitCode=0 Nov 24 01:30:09 crc kubenswrapper[4888]: I1124 01:30:09.441605 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerStarted","Data":"1604bb13f7dd549bfc78866ae70ec8ac9955b3576398710a13be51d99b20a47d"} Nov 24 01:30:11 crc kubenswrapper[4888]: I1124 01:30:11.474625 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerStarted","Data":"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb"} Nov 24 01:30:12 crc kubenswrapper[4888]: I1124 01:30:12.490270 4888 generic.go:334] "Generic (PLEG): container finished" podID="3fb76208-02ad-443b-a757-7deb1d079541" containerID="a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb" exitCode=0 Nov 24 01:30:12 crc kubenswrapper[4888]: I1124 01:30:12.490494 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerDied","Data":"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb"} Nov 24 01:30:13 crc kubenswrapper[4888]: I1124 01:30:13.503031 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerStarted","Data":"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d"} Nov 24 01:30:13 crc kubenswrapper[4888]: I1124 01:30:13.529906 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vdrbf" podStartSLOduration=2.972223114 podStartE2EDuration="6.529883669s" podCreationTimestamp="2025-11-24 01:30:07 +0000 UTC" firstStartedPulling="2025-11-24 01:30:09.444729506 +0000 UTC m=+3912.027413590" lastFinishedPulling="2025-11-24 01:30:13.002390091 +0000 UTC m=+3915.585074145" observedRunningTime="2025-11-24 01:30:13.523462298 +0000 UTC m=+3916.106146362" watchObservedRunningTime="2025-11-24 01:30:13.529883669 +0000 UTC m=+3916.112567733" Nov 24 01:30:14 crc kubenswrapper[4888]: I1124 01:30:14.860479 4888 scope.go:117] "RemoveContainer" containerID="7662f3f310053db964245230730762648b4857c92e564ac9c10e4407e0c35db1" Nov 24 01:30:18 crc kubenswrapper[4888]: I1124 01:30:18.128196 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:18 crc kubenswrapper[4888]: I1124 01:30:18.129037 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:18 crc kubenswrapper[4888]: I1124 01:30:18.196999 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:18 crc kubenswrapper[4888]: I1124 01:30:18.657207 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:18 crc kubenswrapper[4888]: I1124 01:30:18.722423 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:20 crc kubenswrapper[4888]: I1124 01:30:20.604307 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vdrbf" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="registry-server" containerID="cri-o://87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d" gracePeriod=2 Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.105445 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.148216 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities\") pod \"3fb76208-02ad-443b-a757-7deb1d079541\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.148334 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content\") pod \"3fb76208-02ad-443b-a757-7deb1d079541\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.148469 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzdrp\" (UniqueName: \"kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp\") pod \"3fb76208-02ad-443b-a757-7deb1d079541\" (UID: \"3fb76208-02ad-443b-a757-7deb1d079541\") " Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.149652 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities" (OuterVolumeSpecName: "utilities") pod "3fb76208-02ad-443b-a757-7deb1d079541" (UID: "3fb76208-02ad-443b-a757-7deb1d079541"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.155210 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp" (OuterVolumeSpecName: "kube-api-access-jzdrp") pod "3fb76208-02ad-443b-a757-7deb1d079541" (UID: "3fb76208-02ad-443b-a757-7deb1d079541"). InnerVolumeSpecName "kube-api-access-jzdrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.208688 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fb76208-02ad-443b-a757-7deb1d079541" (UID: "3fb76208-02ad-443b-a757-7deb1d079541"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.250619 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.250875 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb76208-02ad-443b-a757-7deb1d079541-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.250885 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzdrp\" (UniqueName: \"kubernetes.io/projected/3fb76208-02ad-443b-a757-7deb1d079541-kube-api-access-jzdrp\") on node \"crc\" DevicePath \"\"" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.618191 4888 generic.go:334] "Generic (PLEG): container finished" podID="3fb76208-02ad-443b-a757-7deb1d079541" containerID="87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d" exitCode=0 Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.618248 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerDied","Data":"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d"} Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.618286 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrbf" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.618322 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrbf" event={"ID":"3fb76208-02ad-443b-a757-7deb1d079541","Type":"ContainerDied","Data":"1604bb13f7dd549bfc78866ae70ec8ac9955b3576398710a13be51d99b20a47d"} Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.618345 4888 scope.go:117] "RemoveContainer" containerID="87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.666474 4888 scope.go:117] "RemoveContainer" containerID="a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.671694 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.689022 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vdrbf"] Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.710283 4888 scope.go:117] "RemoveContainer" containerID="9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.756187 4888 scope.go:117] "RemoveContainer" containerID="87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d" Nov 24 01:30:21 crc kubenswrapper[4888]: E1124 01:30:21.756706 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d\": container with ID starting with 87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d not found: ID does not exist" containerID="87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.756753 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d"} err="failed to get container status \"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d\": rpc error: code = NotFound desc = could not find container \"87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d\": container with ID starting with 87b0588c347cc373ac901e647a42e8ea8a762cc37388b9d94aa21458669f227d not found: ID does not exist" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.756789 4888 scope.go:117] "RemoveContainer" containerID="a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb" Nov 24 01:30:21 crc kubenswrapper[4888]: E1124 01:30:21.757297 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb\": container with ID starting with a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb not found: ID does not exist" containerID="a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.757322 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb"} err="failed to get container status \"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb\": rpc error: code = NotFound desc = could not find container \"a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb\": container with ID starting with a4aa15e823c55c0f3fb07c8894ecd2ca1c03e1274cec50fb09d7e5b361d619fb not found: ID does not exist" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.757342 4888 scope.go:117] "RemoveContainer" containerID="9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59" Nov 24 01:30:21 crc kubenswrapper[4888]: E1124 01:30:21.757704 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59\": container with ID starting with 9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59 not found: ID does not exist" containerID="9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59" Nov 24 01:30:21 crc kubenswrapper[4888]: I1124 01:30:21.757744 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59"} err="failed to get container status \"9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59\": rpc error: code = NotFound desc = could not find container \"9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59\": container with ID starting with 9e122ba1bfc2999d028c38263191202803395d814791eadf7f8af0b2f8df9f59 not found: ID does not exist" Nov 24 01:30:22 crc kubenswrapper[4888]: I1124 01:30:22.272636 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb76208-02ad-443b-a757-7deb1d079541" path="/var/lib/kubelet/pods/3fb76208-02ad-443b-a757-7deb1d079541/volumes" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.969362 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:30:59 crc kubenswrapper[4888]: E1124 01:30:59.970463 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="extract-utilities" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.970481 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="extract-utilities" Nov 24 01:30:59 crc kubenswrapper[4888]: E1124 01:30:59.970510 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="extract-content" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.970518 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="extract-content" Nov 24 01:30:59 crc kubenswrapper[4888]: E1124 01:30:59.970549 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="registry-server" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.970557 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="registry-server" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.971003 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb76208-02ad-443b-a757-7deb1d079541" containerName="registry-server" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.972837 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.981212 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.981280 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.981311 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8rc9\" (UniqueName: \"kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:30:59 crc kubenswrapper[4888]: I1124 01:30:59.988474 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.083355 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.083424 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.083443 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8rc9\" (UniqueName: \"kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.083974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.083993 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.105058 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8rc9\" (UniqueName: \"kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9\") pod \"redhat-marketplace-8gcfk\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.306919 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:00 crc kubenswrapper[4888]: I1124 01:31:00.782669 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:31:01 crc kubenswrapper[4888]: I1124 01:31:01.102497 4888 generic.go:334] "Generic (PLEG): container finished" podID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerID="a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0" exitCode=0 Nov 24 01:31:01 crc kubenswrapper[4888]: I1124 01:31:01.102551 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerDied","Data":"a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0"} Nov 24 01:31:01 crc kubenswrapper[4888]: I1124 01:31:01.102856 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerStarted","Data":"5e06290bba90c3abd72f20d3dddbf2ea99877a61bc9593f935663ecfa6712721"} Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.113460 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerStarted","Data":"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2"} Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.369933 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.372391 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.384509 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.533221 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.533258 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9q4\" (UniqueName: \"kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.533359 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.635774 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.635975 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.636002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9q4\" (UniqueName: \"kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.636264 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.636439 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.659279 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9q4\" (UniqueName: \"kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4\") pod \"community-operators-2vflm\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:02 crc kubenswrapper[4888]: I1124 01:31:02.762430 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:03 crc kubenswrapper[4888]: I1124 01:31:03.126682 4888 generic.go:334] "Generic (PLEG): container finished" podID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerID="453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2" exitCode=0 Nov 24 01:31:03 crc kubenswrapper[4888]: I1124 01:31:03.126845 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerDied","Data":"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2"} Nov 24 01:31:03 crc kubenswrapper[4888]: I1124 01:31:03.343956 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:03 crc kubenswrapper[4888]: W1124 01:31:03.350114 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa49a268_f5c0_4ae1_a67f_2426155f539a.slice/crio-a012f665a2618adeb040adceb384f2d741abc193bdff49e1bd238079a348e583 WatchSource:0}: Error finding container a012f665a2618adeb040adceb384f2d741abc193bdff49e1bd238079a348e583: Status 404 returned error can't find the container with id a012f665a2618adeb040adceb384f2d741abc193bdff49e1bd238079a348e583 Nov 24 01:31:04 crc kubenswrapper[4888]: I1124 01:31:04.137290 4888 generic.go:334] "Generic (PLEG): container finished" podID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerID="139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20" exitCode=0 Nov 24 01:31:04 crc kubenswrapper[4888]: I1124 01:31:04.137975 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerDied","Data":"139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20"} Nov 24 01:31:04 crc kubenswrapper[4888]: I1124 01:31:04.138003 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerStarted","Data":"a012f665a2618adeb040adceb384f2d741abc193bdff49e1bd238079a348e583"} Nov 24 01:31:04 crc kubenswrapper[4888]: I1124 01:31:04.149105 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerStarted","Data":"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa"} Nov 24 01:31:04 crc kubenswrapper[4888]: I1124 01:31:04.181192 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8gcfk" podStartSLOduration=2.717016236 podStartE2EDuration="5.181171936s" podCreationTimestamp="2025-11-24 01:30:59 +0000 UTC" firstStartedPulling="2025-11-24 01:31:01.104542701 +0000 UTC m=+3963.687226765" lastFinishedPulling="2025-11-24 01:31:03.568698421 +0000 UTC m=+3966.151382465" observedRunningTime="2025-11-24 01:31:04.180683702 +0000 UTC m=+3966.763367746" watchObservedRunningTime="2025-11-24 01:31:04.181171936 +0000 UTC m=+3966.763855980" Nov 24 01:31:06 crc kubenswrapper[4888]: I1124 01:31:06.168794 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerStarted","Data":"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000"} Nov 24 01:31:07 crc kubenswrapper[4888]: I1124 01:31:07.185495 4888 generic.go:334] "Generic (PLEG): container finished" podID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerID="b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000" exitCode=0 Nov 24 01:31:07 crc kubenswrapper[4888]: I1124 01:31:07.185565 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerDied","Data":"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000"} Nov 24 01:31:08 crc kubenswrapper[4888]: I1124 01:31:08.200477 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerStarted","Data":"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d"} Nov 24 01:31:08 crc kubenswrapper[4888]: I1124 01:31:08.223715 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2vflm" podStartSLOduration=2.710227016 podStartE2EDuration="6.223696811s" podCreationTimestamp="2025-11-24 01:31:02 +0000 UTC" firstStartedPulling="2025-11-24 01:31:04.140115523 +0000 UTC m=+3966.722799567" lastFinishedPulling="2025-11-24 01:31:07.653585318 +0000 UTC m=+3970.236269362" observedRunningTime="2025-11-24 01:31:08.221104289 +0000 UTC m=+3970.803788343" watchObservedRunningTime="2025-11-24 01:31:08.223696811 +0000 UTC m=+3970.806380855" Nov 24 01:31:10 crc kubenswrapper[4888]: I1124 01:31:10.308071 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:10 crc kubenswrapper[4888]: I1124 01:31:10.308121 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:10 crc kubenswrapper[4888]: I1124 01:31:10.390277 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:11 crc kubenswrapper[4888]: I1124 01:31:11.805154 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:11 crc kubenswrapper[4888]: I1124 01:31:11.851618 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:31:12 crc kubenswrapper[4888]: I1124 01:31:12.763761 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:12 crc kubenswrapper[4888]: I1124 01:31:12.763832 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:12 crc kubenswrapper[4888]: I1124 01:31:12.842992 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.257562 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8gcfk" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="registry-server" containerID="cri-o://77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa" gracePeriod=2 Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.312402 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.813394 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.909535 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities\") pod \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.909659 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8rc9\" (UniqueName: \"kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9\") pod \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.909789 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content\") pod \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\" (UID: \"6a4f4773-c4b4-4d45-89c2-b00106c1b053\") " Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.911269 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities" (OuterVolumeSpecName: "utilities") pod "6a4f4773-c4b4-4d45-89c2-b00106c1b053" (UID: "6a4f4773-c4b4-4d45-89c2-b00106c1b053"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.918164 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9" (OuterVolumeSpecName: "kube-api-access-z8rc9") pod "6a4f4773-c4b4-4d45-89c2-b00106c1b053" (UID: "6a4f4773-c4b4-4d45-89c2-b00106c1b053"). InnerVolumeSpecName "kube-api-access-z8rc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:31:13 crc kubenswrapper[4888]: I1124 01:31:13.928271 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a4f4773-c4b4-4d45-89c2-b00106c1b053" (UID: "6a4f4773-c4b4-4d45-89c2-b00106c1b053"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.012471 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8rc9\" (UniqueName: \"kubernetes.io/projected/6a4f4773-c4b4-4d45-89c2-b00106c1b053-kube-api-access-z8rc9\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.012517 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.012531 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4f4773-c4b4-4d45-89c2-b00106c1b053-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.271793 4888 generic.go:334] "Generic (PLEG): container finished" podID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerID="77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa" exitCode=0 Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.271874 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerDied","Data":"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa"} Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.271943 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gcfk" event={"ID":"6a4f4773-c4b4-4d45-89c2-b00106c1b053","Type":"ContainerDied","Data":"5e06290bba90c3abd72f20d3dddbf2ea99877a61bc9593f935663ecfa6712721"} Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.271949 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gcfk" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.271986 4888 scope.go:117] "RemoveContainer" containerID="77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.304781 4888 scope.go:117] "RemoveContainer" containerID="453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.306531 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.316627 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gcfk"] Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.329413 4888 scope.go:117] "RemoveContainer" containerID="a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.364460 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.399367 4888 scope.go:117] "RemoveContainer" containerID="77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa" Nov 24 01:31:14 crc kubenswrapper[4888]: E1124 01:31:14.406823 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa\": container with ID starting with 77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa not found: ID does not exist" containerID="77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.406870 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa"} err="failed to get container status \"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa\": rpc error: code = NotFound desc = could not find container \"77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa\": container with ID starting with 77166c7c2d69d01cad7bb05b2638de79b87847ae2dc5cec7e5815d4c658961fa not found: ID does not exist" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.406905 4888 scope.go:117] "RemoveContainer" containerID="453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2" Nov 24 01:31:14 crc kubenswrapper[4888]: E1124 01:31:14.407631 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2\": container with ID starting with 453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2 not found: ID does not exist" containerID="453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.407653 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2"} err="failed to get container status \"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2\": rpc error: code = NotFound desc = could not find container \"453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2\": container with ID starting with 453b5d2a2f01c94767b3fdee27d21db6a3a74c30522c49f514f9cb315510d9d2 not found: ID does not exist" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.407689 4888 scope.go:117] "RemoveContainer" containerID="a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0" Nov 24 01:31:14 crc kubenswrapper[4888]: E1124 01:31:14.408095 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0\": container with ID starting with a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0 not found: ID does not exist" containerID="a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0" Nov 24 01:31:14 crc kubenswrapper[4888]: I1124 01:31:14.408127 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0"} err="failed to get container status \"a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0\": rpc error: code = NotFound desc = could not find container \"a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0\": container with ID starting with a21611f51068cb68baf11791f8632392fedd799988c5ffa9c7523efb0cf378c0 not found: ID does not exist" Nov 24 01:31:15 crc kubenswrapper[4888]: I1124 01:31:15.280054 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2vflm" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="registry-server" containerID="cri-o://09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d" gracePeriod=2 Nov 24 01:31:15 crc kubenswrapper[4888]: I1124 01:31:15.870381 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.057726 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content\") pod \"aa49a268-f5c0-4ae1-a67f-2426155f539a\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.058037 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj9q4\" (UniqueName: \"kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4\") pod \"aa49a268-f5c0-4ae1-a67f-2426155f539a\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.058949 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities\") pod \"aa49a268-f5c0-4ae1-a67f-2426155f539a\" (UID: \"aa49a268-f5c0-4ae1-a67f-2426155f539a\") " Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.059651 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities" (OuterVolumeSpecName: "utilities") pod "aa49a268-f5c0-4ae1-a67f-2426155f539a" (UID: "aa49a268-f5c0-4ae1-a67f-2426155f539a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.068496 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4" (OuterVolumeSpecName: "kube-api-access-dj9q4") pod "aa49a268-f5c0-4ae1-a67f-2426155f539a" (UID: "aa49a268-f5c0-4ae1-a67f-2426155f539a"). InnerVolumeSpecName "kube-api-access-dj9q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.117116 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa49a268-f5c0-4ae1-a67f-2426155f539a" (UID: "aa49a268-f5c0-4ae1-a67f-2426155f539a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.161352 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.161380 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj9q4\" (UniqueName: \"kubernetes.io/projected/aa49a268-f5c0-4ae1-a67f-2426155f539a-kube-api-access-dj9q4\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.161391 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49a268-f5c0-4ae1-a67f-2426155f539a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.258839 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" path="/var/lib/kubelet/pods/6a4f4773-c4b4-4d45-89c2-b00106c1b053/volumes" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.294492 4888 generic.go:334] "Generic (PLEG): container finished" podID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerID="09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d" exitCode=0 Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.294539 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerDied","Data":"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d"} Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.294569 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vflm" event={"ID":"aa49a268-f5c0-4ae1-a67f-2426155f539a","Type":"ContainerDied","Data":"a012f665a2618adeb040adceb384f2d741abc193bdff49e1bd238079a348e583"} Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.294592 4888 scope.go:117] "RemoveContainer" containerID="09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.294727 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vflm" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.318282 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.331408 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2vflm"] Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.331484 4888 scope.go:117] "RemoveContainer" containerID="b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.355464 4888 scope.go:117] "RemoveContainer" containerID="139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.407160 4888 scope.go:117] "RemoveContainer" containerID="09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d" Nov 24 01:31:16 crc kubenswrapper[4888]: E1124 01:31:16.407784 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d\": container with ID starting with 09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d not found: ID does not exist" containerID="09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.407834 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d"} err="failed to get container status \"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d\": rpc error: code = NotFound desc = could not find container \"09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d\": container with ID starting with 09e3fc18c2356a3d721a7b8a8e2dc91a3e142e6b081117414f1f0c9bede9f31d not found: ID does not exist" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.407862 4888 scope.go:117] "RemoveContainer" containerID="b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000" Nov 24 01:31:16 crc kubenswrapper[4888]: E1124 01:31:16.408111 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000\": container with ID starting with b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000 not found: ID does not exist" containerID="b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.408141 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000"} err="failed to get container status \"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000\": rpc error: code = NotFound desc = could not find container \"b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000\": container with ID starting with b479ad45f56286ee5a1c223c0f22b9053beebbb9115afb8d0743ac8ad5726000 not found: ID does not exist" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.408161 4888 scope.go:117] "RemoveContainer" containerID="139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20" Nov 24 01:31:16 crc kubenswrapper[4888]: E1124 01:31:16.408439 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20\": container with ID starting with 139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20 not found: ID does not exist" containerID="139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20" Nov 24 01:31:16 crc kubenswrapper[4888]: I1124 01:31:16.408466 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20"} err="failed to get container status \"139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20\": rpc error: code = NotFound desc = could not find container \"139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20\": container with ID starting with 139dec4fa6535a54111704ceb40f0467993a33c584f1c4b8a98164ac23622a20 not found: ID does not exist" Nov 24 01:31:18 crc kubenswrapper[4888]: I1124 01:31:18.259361 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" path="/var/lib/kubelet/pods/aa49a268-f5c0-4ae1-a67f-2426155f539a/volumes" Nov 24 01:31:53 crc kubenswrapper[4888]: I1124 01:31:53.591192 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:31:53 crc kubenswrapper[4888]: I1124 01:31:53.591780 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:32:23 crc kubenswrapper[4888]: I1124 01:32:23.591429 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:32:23 crc kubenswrapper[4888]: I1124 01:32:23.592298 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:32:29 crc kubenswrapper[4888]: I1124 01:32:29.292247 4888 generic.go:334] "Generic (PLEG): container finished" podID="59680387-8e94-4e51-9b44-9c0829e262a6" containerID="feef08808100513068d122b513e7db053d166688a6790804811c018fe692d31e" exitCode=0 Nov 24 01:32:29 crc kubenswrapper[4888]: I1124 01:32:29.292377 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" event={"ID":"59680387-8e94-4e51-9b44-9c0829e262a6","Type":"ContainerDied","Data":"feef08808100513068d122b513e7db053d166688a6790804811c018fe692d31e"} Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.163277 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.312976 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" event={"ID":"59680387-8e94-4e51-9b44-9c0829e262a6","Type":"ContainerDied","Data":"35be167aabfb6cfdefe750243ef233b4faca0f54c9b9536ea986ddd0cd9d67cc"} Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.313040 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35be167aabfb6cfdefe750243ef233b4faca0f54c9b9536ea986ddd0cd9d67cc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.313015 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355140 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355428 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355508 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355546 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355643 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355669 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355699 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355729 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355768 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355794 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.355850 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5w6j\" (UniqueName: \"kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j\") pod \"59680387-8e94-4e51-9b44-9c0829e262a6\" (UID: \"59680387-8e94-4e51-9b44-9c0829e262a6\") " Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.363962 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.378749 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph" (OuterVolumeSpecName: "ceph") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.387178 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j" (OuterVolumeSpecName: "kube-api-access-h5w6j") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "kube-api-access-h5w6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.394946 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.405442 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.406602 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory" (OuterVolumeSpecName: "inventory") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.415793 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.416574 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.418074 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.439697 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc"] Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440175 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440196 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440216 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="extract-content" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440224 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="extract-content" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440246 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="extract-utilities" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440254 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="extract-utilities" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440266 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440273 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440309 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59680387-8e94-4e51-9b44-9c0829e262a6" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440319 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="59680387-8e94-4e51-9b44-9c0829e262a6" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440334 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="extract-content" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440340 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="extract-content" Nov 24 01:32:31 crc kubenswrapper[4888]: E1124 01:32:31.440355 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="extract-utilities" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.440363 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="extract-utilities" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.441988 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="59680387-8e94-4e51-9b44-9c0829e262a6" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.442009 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa49a268-f5c0-4ae1-a67f-2426155f539a" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.442026 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4f4773-c4b4-4d45-89c2-b00106c1b053" containerName="registry-server" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.443413 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.444958 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.448678 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460247 4888 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460307 4888 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460320 4888 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460332 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460343 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460354 4888 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460416 4888 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460429 4888 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/59680387-8e94-4e51-9b44-9c0829e262a6-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460442 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5w6j\" (UniqueName: \"kubernetes.io/projected/59680387-8e94-4e51-9b44-9c0829e262a6-kube-api-access-h5w6j\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.460453 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.461393 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc"] Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.463192 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "59680387-8e94-4e51-9b44-9c0829e262a6" (UID: "59680387-8e94-4e51-9b44-9c0829e262a6"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.562283 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.562679 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.562709 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm4r2\" (UniqueName: \"kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.562936 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.563085 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.563138 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.563169 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.563248 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.563360 4888 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/59680387-8e94-4e51-9b44-9c0829e262a6-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665134 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665219 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665272 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665298 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm4r2\" (UniqueName: \"kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665375 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665418 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665440 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.665458 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.668354 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.668613 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.669286 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.669282 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.669455 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.678388 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.678831 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.680903 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm4r2\" (UniqueName: \"kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t54vc\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:31 crc kubenswrapper[4888]: I1124 01:32:31.856528 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:32:32 crc kubenswrapper[4888]: I1124 01:32:32.379421 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc"] Nov 24 01:32:32 crc kubenswrapper[4888]: W1124 01:32:32.394755 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7009799b_7622_43c3_856d_7a368de7608f.slice/crio-7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04 WatchSource:0}: Error finding container 7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04: Status 404 returned error can't find the container with id 7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04 Nov 24 01:32:33 crc kubenswrapper[4888]: I1124 01:32:33.333571 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" event={"ID":"7009799b-7622-43c3-856d-7a368de7608f","Type":"ContainerStarted","Data":"332031b4d98e1a22a93da6f26da070deae295380fec7cac23caddc6f446e9f03"} Nov 24 01:32:33 crc kubenswrapper[4888]: I1124 01:32:33.334379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" event={"ID":"7009799b-7622-43c3-856d-7a368de7608f","Type":"ContainerStarted","Data":"7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04"} Nov 24 01:32:33 crc kubenswrapper[4888]: I1124 01:32:33.368079 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" podStartSLOduration=1.792093165 podStartE2EDuration="2.368057164s" podCreationTimestamp="2025-11-24 01:32:31 +0000 UTC" firstStartedPulling="2025-11-24 01:32:32.407133241 +0000 UTC m=+4054.989817295" lastFinishedPulling="2025-11-24 01:32:32.98309725 +0000 UTC m=+4055.565781294" observedRunningTime="2025-11-24 01:32:33.355874452 +0000 UTC m=+4055.938558516" watchObservedRunningTime="2025-11-24 01:32:33.368057164 +0000 UTC m=+4055.950741218" Nov 24 01:32:53 crc kubenswrapper[4888]: I1124 01:32:53.592082 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:32:53 crc kubenswrapper[4888]: I1124 01:32:53.593173 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:32:53 crc kubenswrapper[4888]: I1124 01:32:53.593282 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:32:53 crc kubenswrapper[4888]: I1124 01:32:53.594887 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:32:53 crc kubenswrapper[4888]: I1124 01:32:53.595005 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf" gracePeriod=600 Nov 24 01:32:54 crc kubenswrapper[4888]: I1124 01:32:54.571876 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf" exitCode=0 Nov 24 01:32:54 crc kubenswrapper[4888]: I1124 01:32:54.571967 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf"} Nov 24 01:32:54 crc kubenswrapper[4888]: I1124 01:32:54.572586 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829"} Nov 24 01:32:54 crc kubenswrapper[4888]: I1124 01:32:54.572613 4888 scope.go:117] "RemoveContainer" containerID="0942b2decf762fb9389458b6a9062449e4c9960ae1e1f2588298a59b0d75d589" Nov 24 01:34:53 crc kubenswrapper[4888]: I1124 01:34:53.591758 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:34:53 crc kubenswrapper[4888]: I1124 01:34:53.593890 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:35:23 crc kubenswrapper[4888]: I1124 01:35:23.592415 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:35:23 crc kubenswrapper[4888]: I1124 01:35:23.593219 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:35:46 crc kubenswrapper[4888]: I1124 01:35:46.823932 4888 generic.go:334] "Generic (PLEG): container finished" podID="7009799b-7622-43c3-856d-7a368de7608f" containerID="332031b4d98e1a22a93da6f26da070deae295380fec7cac23caddc6f446e9f03" exitCode=0 Nov 24 01:35:46 crc kubenswrapper[4888]: I1124 01:35:46.824083 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" event={"ID":"7009799b-7622-43c3-856d-7a368de7608f","Type":"ContainerDied","Data":"332031b4d98e1a22a93da6f26da070deae295380fec7cac23caddc6f446e9f03"} Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.413102 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548039 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548088 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548183 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548293 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548322 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548346 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548432 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm4r2\" (UniqueName: \"kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.548526 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key\") pod \"7009799b-7622-43c3-856d-7a368de7608f\" (UID: \"7009799b-7622-43c3-856d-7a368de7608f\") " Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.555748 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2" (OuterVolumeSpecName: "kube-api-access-zm4r2") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "kube-api-access-zm4r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.565787 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.565928 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph" (OuterVolumeSpecName: "ceph") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.587404 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.596578 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.598057 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.598545 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.603136 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory" (OuterVolumeSpecName: "inventory") pod "7009799b-7622-43c3-856d-7a368de7608f" (UID: "7009799b-7622-43c3-856d-7a368de7608f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650616 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650650 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650665 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm4r2\" (UniqueName: \"kubernetes.io/projected/7009799b-7622-43c3-856d-7a368de7608f-kube-api-access-zm4r2\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650677 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650690 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650700 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650710 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.650722 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7009799b-7622-43c3-856d-7a368de7608f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.851143 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" event={"ID":"7009799b-7622-43c3-856d-7a368de7608f","Type":"ContainerDied","Data":"7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04"} Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.851184 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c205c63d3ae313d040fc710a5f27f06d86ca32ef697de2e958233cd39d01d04" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.851222 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t54vc" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.962939 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9"] Nov 24 01:35:48 crc kubenswrapper[4888]: E1124 01:35:48.963465 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7009799b-7622-43c3-856d-7a368de7608f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.963482 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7009799b-7622-43c3-856d-7a368de7608f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.963689 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7009799b-7622-43c3-856d-7a368de7608f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.964612 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.966581 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.967188 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.967267 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.967198 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.967401 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.969747 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:35:48 crc kubenswrapper[4888]: I1124 01:35:48.975176 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9"] Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058502 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058567 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058589 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058609 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hmpq\" (UniqueName: \"kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058656 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058686 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058722 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.058831 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160515 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160613 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160644 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160672 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hmpq\" (UniqueName: \"kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160748 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160798 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.160895 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.161016 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.447015 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.449188 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.449441 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.449807 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.450485 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.451040 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.452103 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.453053 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hmpq\" (UniqueName: \"kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:49 crc kubenswrapper[4888]: I1124 01:35:49.581970 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:35:50 crc kubenswrapper[4888]: I1124 01:35:50.288158 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9"] Nov 24 01:35:50 crc kubenswrapper[4888]: I1124 01:35:50.292481 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:35:50 crc kubenswrapper[4888]: I1124 01:35:50.876022 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" event={"ID":"1fb69384-3ad1-424b-93ad-2716c502bc4d","Type":"ContainerStarted","Data":"f6eb91111e3cdec8ea923b1c3695841ea6376961f6853a4ccf09e5f59ccd1da4"} Nov 24 01:35:52 crc kubenswrapper[4888]: I1124 01:35:52.902744 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" event={"ID":"1fb69384-3ad1-424b-93ad-2716c502bc4d","Type":"ContainerStarted","Data":"862b703d132ba3047404dc7fcce20e622242d041d3dbea3b5632a1d30cee0fc4"} Nov 24 01:35:52 crc kubenswrapper[4888]: I1124 01:35:52.919829 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" podStartSLOduration=4.46536514 podStartE2EDuration="4.919797199s" podCreationTimestamp="2025-11-24 01:35:48 +0000 UTC" firstStartedPulling="2025-11-24 01:35:50.292286855 +0000 UTC m=+4252.874970899" lastFinishedPulling="2025-11-24 01:35:50.746718864 +0000 UTC m=+4253.329402958" observedRunningTime="2025-11-24 01:35:52.91912379 +0000 UTC m=+4255.501807834" watchObservedRunningTime="2025-11-24 01:35:52.919797199 +0000 UTC m=+4255.502481243" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.591917 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.592537 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.592602 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.593798 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.593929 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" gracePeriod=600 Nov 24 01:35:53 crc kubenswrapper[4888]: E1124 01:35:53.758003 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.922949 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" exitCode=0 Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.923074 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829"} Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.923171 4888 scope.go:117] "RemoveContainer" containerID="006fb10069a7313c4cdad2e233ff9bfcbc64cff3ad30b78cd9f950d1e88499bf" Nov 24 01:35:53 crc kubenswrapper[4888]: I1124 01:35:53.924088 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:35:53 crc kubenswrapper[4888]: E1124 01:35:53.924533 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:36:08 crc kubenswrapper[4888]: I1124 01:36:08.252436 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:36:08 crc kubenswrapper[4888]: E1124 01:36:08.253402 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:36:23 crc kubenswrapper[4888]: I1124 01:36:23.244779 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:36:23 crc kubenswrapper[4888]: E1124 01:36:23.245552 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:36:38 crc kubenswrapper[4888]: I1124 01:36:38.251210 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:36:38 crc kubenswrapper[4888]: E1124 01:36:38.251923 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.581440 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.585103 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.609076 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.647418 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.647610 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxmsz\" (UniqueName: \"kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.647651 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.749674 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.749925 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxmsz\" (UniqueName: \"kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.749982 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.750140 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.750527 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.774269 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxmsz\" (UniqueName: \"kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz\") pod \"redhat-operators-dgtbn\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:44 crc kubenswrapper[4888]: I1124 01:36:44.913911 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:36:45 crc kubenswrapper[4888]: I1124 01:36:45.386398 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:36:45 crc kubenswrapper[4888]: I1124 01:36:45.539108 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerStarted","Data":"47cb6892fef9ef2e7d17d6546fb085b5eb5141d654cca533d315534aba6ecd73"} Nov 24 01:36:46 crc kubenswrapper[4888]: I1124 01:36:46.549099 4888 generic.go:334] "Generic (PLEG): container finished" podID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerID="cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4" exitCode=0 Nov 24 01:36:46 crc kubenswrapper[4888]: I1124 01:36:46.549204 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerDied","Data":"cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4"} Nov 24 01:36:48 crc kubenswrapper[4888]: I1124 01:36:48.571697 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerStarted","Data":"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251"} Nov 24 01:36:53 crc kubenswrapper[4888]: I1124 01:36:53.245077 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:36:53 crc kubenswrapper[4888]: E1124 01:36:53.245804 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:36:54 crc kubenswrapper[4888]: I1124 01:36:54.641670 4888 generic.go:334] "Generic (PLEG): container finished" podID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerID="b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251" exitCode=0 Nov 24 01:36:54 crc kubenswrapper[4888]: I1124 01:36:54.641751 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerDied","Data":"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251"} Nov 24 01:36:55 crc kubenswrapper[4888]: I1124 01:36:55.654006 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerStarted","Data":"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2"} Nov 24 01:36:55 crc kubenswrapper[4888]: I1124 01:36:55.684214 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dgtbn" podStartSLOduration=3.172541421 podStartE2EDuration="11.684189666s" podCreationTimestamp="2025-11-24 01:36:44 +0000 UTC" firstStartedPulling="2025-11-24 01:36:46.564790417 +0000 UTC m=+4309.147474451" lastFinishedPulling="2025-11-24 01:36:55.076438612 +0000 UTC m=+4317.659122696" observedRunningTime="2025-11-24 01:36:55.671525381 +0000 UTC m=+4318.254209435" watchObservedRunningTime="2025-11-24 01:36:55.684189666 +0000 UTC m=+4318.266873720" Nov 24 01:37:04 crc kubenswrapper[4888]: I1124 01:37:04.914785 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:04 crc kubenswrapper[4888]: I1124 01:37:04.915646 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:05 crc kubenswrapper[4888]: I1124 01:37:05.400349 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:05 crc kubenswrapper[4888]: I1124 01:37:05.831229 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:05 crc kubenswrapper[4888]: I1124 01:37:05.886741 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:37:07 crc kubenswrapper[4888]: I1124 01:37:07.798806 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dgtbn" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="registry-server" containerID="cri-o://c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2" gracePeriod=2 Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.251788 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:37:08 crc kubenswrapper[4888]: E1124 01:37:08.252438 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.318304 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.409737 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content\") pod \"574a6c31-5202-4265-b169-e9fbcf8c3756\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.410249 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxmsz\" (UniqueName: \"kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz\") pod \"574a6c31-5202-4265-b169-e9fbcf8c3756\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.410298 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities\") pod \"574a6c31-5202-4265-b169-e9fbcf8c3756\" (UID: \"574a6c31-5202-4265-b169-e9fbcf8c3756\") " Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.411297 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities" (OuterVolumeSpecName: "utilities") pod "574a6c31-5202-4265-b169-e9fbcf8c3756" (UID: "574a6c31-5202-4265-b169-e9fbcf8c3756"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.416022 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz" (OuterVolumeSpecName: "kube-api-access-zxmsz") pod "574a6c31-5202-4265-b169-e9fbcf8c3756" (UID: "574a6c31-5202-4265-b169-e9fbcf8c3756"). InnerVolumeSpecName "kube-api-access-zxmsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.506765 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "574a6c31-5202-4265-b169-e9fbcf8c3756" (UID: "574a6c31-5202-4265-b169-e9fbcf8c3756"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.513173 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.513202 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxmsz\" (UniqueName: \"kubernetes.io/projected/574a6c31-5202-4265-b169-e9fbcf8c3756-kube-api-access-zxmsz\") on node \"crc\" DevicePath \"\"" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.513212 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574a6c31-5202-4265-b169-e9fbcf8c3756-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.814268 4888 generic.go:334] "Generic (PLEG): container finished" podID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerID="c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2" exitCode=0 Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.814322 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerDied","Data":"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2"} Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.814422 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dgtbn" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.815128 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dgtbn" event={"ID":"574a6c31-5202-4265-b169-e9fbcf8c3756","Type":"ContainerDied","Data":"47cb6892fef9ef2e7d17d6546fb085b5eb5141d654cca533d315534aba6ecd73"} Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.815192 4888 scope.go:117] "RemoveContainer" containerID="c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.842853 4888 scope.go:117] "RemoveContainer" containerID="b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.872727 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.888205 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dgtbn"] Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.897275 4888 scope.go:117] "RemoveContainer" containerID="cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.951034 4888 scope.go:117] "RemoveContainer" containerID="c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2" Nov 24 01:37:08 crc kubenswrapper[4888]: E1124 01:37:08.951741 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2\": container with ID starting with c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2 not found: ID does not exist" containerID="c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.951780 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2"} err="failed to get container status \"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2\": rpc error: code = NotFound desc = could not find container \"c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2\": container with ID starting with c2667c4104320bc06dc9800400cc9cb6766490dddd120c2efeb807def6fd56a2 not found: ID does not exist" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.951847 4888 scope.go:117] "RemoveContainer" containerID="b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251" Nov 24 01:37:08 crc kubenswrapper[4888]: E1124 01:37:08.952352 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251\": container with ID starting with b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251 not found: ID does not exist" containerID="b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.952530 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251"} err="failed to get container status \"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251\": rpc error: code = NotFound desc = could not find container \"b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251\": container with ID starting with b3cbdde9702570092ed5632e0e7b3eaa58fea0516c8a49906581c4c450bf9251 not found: ID does not exist" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.952649 4888 scope.go:117] "RemoveContainer" containerID="cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4" Nov 24 01:37:08 crc kubenswrapper[4888]: E1124 01:37:08.953088 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4\": container with ID starting with cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4 not found: ID does not exist" containerID="cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4" Nov 24 01:37:08 crc kubenswrapper[4888]: I1124 01:37:08.953127 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4"} err="failed to get container status \"cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4\": rpc error: code = NotFound desc = could not find container \"cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4\": container with ID starting with cba6df729bf12aa090fcd2c83c4749351bdf50ba9d5055c00a06a6d04cffdfe4 not found: ID does not exist" Nov 24 01:37:10 crc kubenswrapper[4888]: I1124 01:37:10.267579 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" path="/var/lib/kubelet/pods/574a6c31-5202-4265-b169-e9fbcf8c3756/volumes" Nov 24 01:37:20 crc kubenswrapper[4888]: I1124 01:37:20.246536 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:37:20 crc kubenswrapper[4888]: E1124 01:37:20.247612 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:37:32 crc kubenswrapper[4888]: I1124 01:37:32.246522 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:37:32 crc kubenswrapper[4888]: E1124 01:37:32.248885 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:37:44 crc kubenswrapper[4888]: I1124 01:37:44.246181 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:37:44 crc kubenswrapper[4888]: E1124 01:37:44.247143 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:37:56 crc kubenswrapper[4888]: I1124 01:37:56.245290 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:37:56 crc kubenswrapper[4888]: E1124 01:37:56.245952 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:38:09 crc kubenswrapper[4888]: I1124 01:38:09.245401 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:38:09 crc kubenswrapper[4888]: E1124 01:38:09.246248 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:38:21 crc kubenswrapper[4888]: I1124 01:38:21.246659 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:38:21 crc kubenswrapper[4888]: E1124 01:38:21.247636 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:38:22 crc kubenswrapper[4888]: I1124 01:38:22.882287 4888 generic.go:334] "Generic (PLEG): container finished" podID="1fb69384-3ad1-424b-93ad-2716c502bc4d" containerID="862b703d132ba3047404dc7fcce20e622242d041d3dbea3b5632a1d30cee0fc4" exitCode=0 Nov 24 01:38:22 crc kubenswrapper[4888]: I1124 01:38:22.882605 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" event={"ID":"1fb69384-3ad1-424b-93ad-2716c502bc4d","Type":"ContainerDied","Data":"862b703d132ba3047404dc7fcce20e622242d041d3dbea3b5632a1d30cee0fc4"} Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.458805 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.625599 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hmpq\" (UniqueName: \"kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.625735 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.625858 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.625940 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.626077 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.626171 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.626343 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.626476 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph\") pod \"1fb69384-3ad1-424b-93ad-2716c502bc4d\" (UID: \"1fb69384-3ad1-424b-93ad-2716c502bc4d\") " Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.632942 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq" (OuterVolumeSpecName: "kube-api-access-4hmpq") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "kube-api-access-4hmpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.634273 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph" (OuterVolumeSpecName: "ceph") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.646772 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.667332 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.668418 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.675263 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.684361 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.689917 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory" (OuterVolumeSpecName: "inventory") pod "1fb69384-3ad1-424b-93ad-2716c502bc4d" (UID: "1fb69384-3ad1-424b-93ad-2716c502bc4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.729913 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.729967 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.729987 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.730006 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.730025 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hmpq\" (UniqueName: \"kubernetes.io/projected/1fb69384-3ad1-424b-93ad-2716c502bc4d-kube-api-access-4hmpq\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.730043 4888 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.730061 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.730080 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/1fb69384-3ad1-424b-93ad-2716c502bc4d-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.919191 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" event={"ID":"1fb69384-3ad1-424b-93ad-2716c502bc4d","Type":"ContainerDied","Data":"f6eb91111e3cdec8ea923b1c3695841ea6376961f6853a4ccf09e5f59ccd1da4"} Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.919236 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6eb91111e3cdec8ea923b1c3695841ea6376961f6853a4ccf09e5f59ccd1da4" Nov 24 01:38:24 crc kubenswrapper[4888]: I1124 01:38:24.919289 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.054337 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf"] Nov 24 01:38:25 crc kubenswrapper[4888]: E1124 01:38:25.055008 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="registry-server" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055032 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="registry-server" Nov 24 01:38:25 crc kubenswrapper[4888]: E1124 01:38:25.055104 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="extract-utilities" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055114 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="extract-utilities" Nov 24 01:38:25 crc kubenswrapper[4888]: E1124 01:38:25.055169 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb69384-3ad1-424b-93ad-2716c502bc4d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055179 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb69384-3ad1-424b-93ad-2716c502bc4d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:25 crc kubenswrapper[4888]: E1124 01:38:25.055196 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="extract-content" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055206 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="extract-content" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055533 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb69384-3ad1-424b-93ad-2716c502bc4d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.055872 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="574a6c31-5202-4265-b169-e9fbcf8c3756" containerName="registry-server" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.057177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.059046 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7f2jq" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.062099 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.062255 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.062523 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.062791 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.066706 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.076898 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf"] Nov 24 01:38:25 crc kubenswrapper[4888]: E1124 01:38:25.220723 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fb69384_3ad1_424b_93ad_2716c502bc4d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fb69384_3ad1_424b_93ad_2716c502bc4d.slice/crio-f6eb91111e3cdec8ea923b1c3695841ea6376961f6853a4ccf09e5f59ccd1da4\": RecentStats: unable to find data in memory cache]" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240573 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240641 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240680 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdv42\" (UniqueName: \"kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240778 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240824 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.240846 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.342600 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.342701 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.342889 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.343005 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.343111 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdv42\" (UniqueName: \"kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.343554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.348278 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.348965 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.349091 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.350132 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.350937 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.376332 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdv42\" (UniqueName: \"kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cvzdf\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:25 crc kubenswrapper[4888]: I1124 01:38:25.397287 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:26 crc kubenswrapper[4888]: I1124 01:38:26.032487 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf"] Nov 24 01:38:26 crc kubenswrapper[4888]: I1124 01:38:26.944472 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" event={"ID":"91206204-242c-4c3d-b686-9a4beebb76f1","Type":"ContainerStarted","Data":"3ce72cd4f158358cee6f20c7c9f1f3026b54870cf22740080fca3a7138a1c755"} Nov 24 01:38:26 crc kubenswrapper[4888]: I1124 01:38:26.944977 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" event={"ID":"91206204-242c-4c3d-b686-9a4beebb76f1","Type":"ContainerStarted","Data":"01b0964ba2bda609892fcfd5e29fa823ca7e3a0372755ab153464cf6f56d9a32"} Nov 24 01:38:26 crc kubenswrapper[4888]: I1124 01:38:26.965554 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" podStartSLOduration=1.561219245 podStartE2EDuration="1.96553024s" podCreationTimestamp="2025-11-24 01:38:25 +0000 UTC" firstStartedPulling="2025-11-24 01:38:26.040415176 +0000 UTC m=+4408.623099220" lastFinishedPulling="2025-11-24 01:38:26.444726171 +0000 UTC m=+4409.027410215" observedRunningTime="2025-11-24 01:38:26.959252885 +0000 UTC m=+4409.541936919" watchObservedRunningTime="2025-11-24 01:38:26.96553024 +0000 UTC m=+4409.548214314" Nov 24 01:38:34 crc kubenswrapper[4888]: I1124 01:38:34.245309 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:38:34 crc kubenswrapper[4888]: E1124 01:38:34.246284 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:38:41 crc kubenswrapper[4888]: I1124 01:38:41.100056 4888 generic.go:334] "Generic (PLEG): container finished" podID="91206204-242c-4c3d-b686-9a4beebb76f1" containerID="3ce72cd4f158358cee6f20c7c9f1f3026b54870cf22740080fca3a7138a1c755" exitCode=0 Nov 24 01:38:41 crc kubenswrapper[4888]: I1124 01:38:41.100117 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" event={"ID":"91206204-242c-4c3d-b686-9a4beebb76f1","Type":"ContainerDied","Data":"3ce72cd4f158358cee6f20c7c9f1f3026b54870cf22740080fca3a7138a1c755"} Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.653130 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.729599 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdv42\" (UniqueName: \"kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.729691 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.729782 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.730017 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.730105 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.730202 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1\") pod \"91206204-242c-4c3d-b686-9a4beebb76f1\" (UID: \"91206204-242c-4c3d-b686-9a4beebb76f1\") " Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.734726 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42" (OuterVolumeSpecName: "kube-api-access-mdv42") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "kube-api-access-mdv42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.749149 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph" (OuterVolumeSpecName: "ceph") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.762893 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.766076 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory" (OuterVolumeSpecName: "inventory") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.773751 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.785679 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "91206204-242c-4c3d-b686-9a4beebb76f1" (UID: "91206204-242c-4c3d-b686-9a4beebb76f1"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833442 4888 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833495 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833515 4888 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833529 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdv42\" (UniqueName: \"kubernetes.io/projected/91206204-242c-4c3d-b686-9a4beebb76f1-kube-api-access-mdv42\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833542 4888 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:42 crc kubenswrapper[4888]: I1124 01:38:42.833555 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91206204-242c-4c3d-b686-9a4beebb76f1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:38:43 crc kubenswrapper[4888]: I1124 01:38:43.128172 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" event={"ID":"91206204-242c-4c3d-b686-9a4beebb76f1","Type":"ContainerDied","Data":"01b0964ba2bda609892fcfd5e29fa823ca7e3a0372755ab153464cf6f56d9a32"} Nov 24 01:38:43 crc kubenswrapper[4888]: I1124 01:38:43.128969 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01b0964ba2bda609892fcfd5e29fa823ca7e3a0372755ab153464cf6f56d9a32" Nov 24 01:38:43 crc kubenswrapper[4888]: I1124 01:38:43.128261 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cvzdf" Nov 24 01:38:49 crc kubenswrapper[4888]: I1124 01:38:49.245587 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:38:49 crc kubenswrapper[4888]: E1124 01:38:49.246485 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.505804 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 01:38:59 crc kubenswrapper[4888]: E1124 01:38:59.506680 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91206204-242c-4c3d-b686-9a4beebb76f1" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.506697 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="91206204-242c-4c3d-b686-9a4beebb76f1" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.506929 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="91206204-242c-4c3d-b686-9a4beebb76f1" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.508022 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.514535 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.517783 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.520073 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.589118 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.591610 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.595482 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.607421 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635230 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635288 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635335 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635351 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-run\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635430 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635447 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635474 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635492 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635520 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635538 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635639 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635678 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pkr\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-kube-api-access-s2pkr\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635759 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635868 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.635911 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.636022 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738053 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738097 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pkr\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-kube-api-access-s2pkr\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738128 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-dev\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738151 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738178 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738198 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738222 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-run\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738245 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738265 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738294 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-ceph\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738315 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738330 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738358 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738374 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738399 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738417 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738433 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-run\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738451 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738475 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-lib-modules\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738501 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738525 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738542 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z66gx\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-kube-api-access-z66gx\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738561 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738581 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738597 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738613 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-scripts\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738629 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-sys\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738646 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738666 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738700 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738716 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.738733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.739756 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.739802 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740149 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740177 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740430 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740528 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740509 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740800 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.740943 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-run\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.741190 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a2644986-db7d-4368-956b-65dc26585e3b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.840002 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.840064 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.840222 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.840683 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.840100 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841049 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-lib-modules\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841103 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841156 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z66gx\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-kube-api-access-z66gx\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841170 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-lib-modules\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841216 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841236 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841262 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-scripts\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841288 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-sys\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841411 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841490 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-dev\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841512 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841520 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841555 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-sys\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841831 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-dev\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841924 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-run\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.841960 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.842017 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-ceph\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.842043 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.842155 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.842200 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-run\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:38:59 crc kubenswrapper[4888]: I1124 01:38:59.842320 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a08f210-f6e8-46f3-9796-8f4236a53dd1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.045204 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.045598 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.045934 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.046465 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.046952 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2644986-db7d-4368-956b-65dc26585e3b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.048048 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pkr\" (UniqueName: \"kubernetes.io/projected/a2644986-db7d-4368-956b-65dc26585e3b-kube-api-access-s2pkr\") pod \"cinder-volume-volume1-0\" (UID: \"a2644986-db7d-4368-956b-65dc26585e3b\") " pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.048194 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-scripts\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.049452 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.049682 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.050888 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-ceph\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.051612 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z66gx\" (UniqueName: \"kubernetes.io/projected/5a08f210-f6e8-46f3-9796-8f4236a53dd1-kube-api-access-z66gx\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.051624 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a08f210-f6e8-46f3-9796-8f4236a53dd1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5a08f210-f6e8-46f3-9796-8f4236a53dd1\") " pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.134754 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.210138 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.321026 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.323383 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.328096 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.334058 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-sjbj7" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.334341 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.334446 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.411871 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.459880 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.459932 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfqwm\" (UniqueName: \"kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.459969 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.460069 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.460090 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.489837 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.492447 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.509191 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.509339 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.509411 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.509446 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5v4qp" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.516888 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.559776 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.563734 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.566673 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.575109 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.577065 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfqwm\" (UniqueName: \"kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.577128 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.577246 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.577273 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.577400 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.580884 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.582265 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.583023 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.583059 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.584864 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.593328 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.596431 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-4j9db"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.600584 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: E1124 01:39:00.605239 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-cdhrf logs scripts], unattached volumes=[], failed to process volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-cdhrf logs scripts]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="f2d62516-a39b-4748-944c-dbf1a52e2980" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.605494 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-4j9db"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.613415 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfqwm\" (UniqueName: \"kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm\") pod \"horizon-66644f99b9-l6svl\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.630598 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.642285 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.649918 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.678982 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679024 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679051 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679072 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679091 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679107 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679177 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679192 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679214 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679257 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679276 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679301 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679319 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n74pf\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679345 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdhrf\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679362 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679384 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679424 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.679569 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.692913 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-129e-account-create-8fkkk"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.712195 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.715883 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.750897 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-129e-account-create-8fkkk"] Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.778859 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781051 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781131 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781166 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781185 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtczv\" (UniqueName: \"kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781258 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781382 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781409 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781435 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781488 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781511 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781555 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7dvj\" (UniqueName: \"kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781579 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781597 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781639 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781695 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781718 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781733 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781783 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781805 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781880 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr5f5\" (UniqueName: \"kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.781923 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782434 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782459 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782485 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782505 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n74pf\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782535 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdhrf\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782552 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.783208 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782108 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.782390 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.784401 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.785073 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.786241 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.793460 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.794170 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.798377 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.798872 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.798937 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.799251 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.801143 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.801351 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.801366 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.802561 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.804836 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdhrf\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.813284 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n74pf\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.838767 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.880214 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885191 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885248 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtczv\" (UniqueName: \"kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885298 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885322 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7dvj\" (UniqueName: \"kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885375 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885412 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885441 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885490 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr5f5\" (UniqueName: \"kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885518 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.885877 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.886157 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.886240 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.886542 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.887286 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.899298 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.910080 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7dvj\" (UniqueName: \"kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj\") pod \"manila-129e-account-create-8fkkk\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.912626 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr5f5\" (UniqueName: \"kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5\") pod \"horizon-5d8d9c897-jxm7k\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.913021 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtczv\" (UniqueName: \"kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv\") pod \"manila-db-create-4j9db\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.940343 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4j9db" Nov 24 01:39:00 crc kubenswrapper[4888]: I1124 01:39:00.986610 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.049426 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.139195 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.248759 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:39:01 crc kubenswrapper[4888]: E1124 01:39:01.249299 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.305946 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.369583 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.405588 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.406418 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:01 crc kubenswrapper[4888]: W1124 01:39:01.415904 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99da32f4_1f0b_4d1b_9456_63c4df848427.slice/crio-011fc40162721804cb934c5b76405a6e8a41759158e63facabbfd35a6b4fa0ef WatchSource:0}: Error finding container 011fc40162721804cb934c5b76405a6e8a41759158e63facabbfd35a6b4fa0ef: Status 404 returned error can't find the container with id 011fc40162721804cb934c5b76405a6e8a41759158e63facabbfd35a6b4fa0ef Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.474937 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 01:39:01 crc kubenswrapper[4888]: W1124 01:39:01.501153 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a08f210_f6e8_46f3_9796_8f4236a53dd1.slice/crio-a5d6c0b0553f2bc2f63e4d517d9eb74b9d7c3190dd8e92412896f45e700e2c6f WatchSource:0}: Error finding container a5d6c0b0553f2bc2f63e4d517d9eb74b9d7c3190dd8e92412896f45e700e2c6f: Status 404 returned error can't find the container with id a5d6c0b0553f2bc2f63e4d517d9eb74b9d7c3190dd8e92412896f45e700e2c6f Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506438 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506539 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506609 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506656 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506691 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506707 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506771 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdhrf\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506919 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.506942 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f2d62516-a39b-4748-944c-dbf1a52e2980\" (UID: \"f2d62516-a39b-4748-944c-dbf1a52e2980\") " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.507434 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs" (OuterVolumeSpecName: "logs") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.508349 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.514744 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.514916 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.516269 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.518048 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data" (OuterVolumeSpecName: "config-data") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.518146 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts" (OuterVolumeSpecName: "scripts") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.518339 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph" (OuterVolumeSpecName: "ceph") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.519125 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf" (OuterVolumeSpecName: "kube-api-access-cdhrf") pod "f2d62516-a39b-4748-944c-dbf1a52e2980" (UID: "f2d62516-a39b-4748-944c-dbf1a52e2980"). InnerVolumeSpecName "kube-api-access-cdhrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610228 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610281 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610293 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610301 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610309 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610317 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610324 4888 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d62516-a39b-4748-944c-dbf1a52e2980-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610334 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d62516-a39b-4748-944c-dbf1a52e2980-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.610342 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdhrf\" (UniqueName: \"kubernetes.io/projected/f2d62516-a39b-4748-944c-dbf1a52e2980-kube-api-access-cdhrf\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.636483 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.712756 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.777243 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-4j9db"] Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.876520 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:01 crc kubenswrapper[4888]: W1124 01:39:01.877398 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce8d408_043a_42be_8d65_a9a96c78670a.slice/crio-d22ad00a23030c923813d4eb06f0cb13f25b769aa77a079eac685fa6c70f575f WatchSource:0}: Error finding container d22ad00a23030c923813d4eb06f0cb13f25b769aa77a079eac685fa6c70f575f: Status 404 returned error can't find the container with id d22ad00a23030c923813d4eb06f0cb13f25b769aa77a079eac685fa6c70f575f Nov 24 01:39:01 crc kubenswrapper[4888]: I1124 01:39:01.929096 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-129e-account-create-8fkkk"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.017248 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.379554 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerStarted","Data":"78d5139d8bfaf12f30193b84469beeeb8fa7704ebae007c3655facb4a748ac76"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.381168 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerStarted","Data":"011fc40162721804cb934c5b76405a6e8a41759158e63facabbfd35a6b4fa0ef"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.386913 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerStarted","Data":"d22ad00a23030c923813d4eb06f0cb13f25b769aa77a079eac685fa6c70f575f"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.390554 4888 generic.go:334] "Generic (PLEG): container finished" podID="3a4de536-bef6-4f2b-ac46-64a43f970422" containerID="5eda77f7cea7914596e6924e60779310c774daad78688d70a610ac3ba1e32049" exitCode=0 Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.390602 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4j9db" event={"ID":"3a4de536-bef6-4f2b-ac46-64a43f970422","Type":"ContainerDied","Data":"5eda77f7cea7914596e6924e60779310c774daad78688d70a610ac3ba1e32049"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.390618 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4j9db" event={"ID":"3a4de536-bef6-4f2b-ac46-64a43f970422","Type":"ContainerStarted","Data":"635eec9e1ac134afdd1c158481f43d6088488286b26edbdac275c5b56723b1cc"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.392947 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5a08f210-f6e8-46f3-9796-8f4236a53dd1","Type":"ContainerStarted","Data":"a5d6c0b0553f2bc2f63e4d517d9eb74b9d7c3190dd8e92412896f45e700e2c6f"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.394830 4888 generic.go:334] "Generic (PLEG): container finished" podID="76da0164-f26f-445f-bc36-401a45149229" containerID="55e094c08ed9c3977990f5c6d7782536d16892a8ba97a266219b2d15611b3014" exitCode=0 Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.394898 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-129e-account-create-8fkkk" event={"ID":"76da0164-f26f-445f-bc36-401a45149229","Type":"ContainerDied","Data":"55e094c08ed9c3977990f5c6d7782536d16892a8ba97a266219b2d15611b3014"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.394920 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-129e-account-create-8fkkk" event={"ID":"76da0164-f26f-445f-bc36-401a45149229","Type":"ContainerStarted","Data":"69f077c2bf3ade7c2db8f0b59ea47b9e73f4d16b13ce4494fa0c476fd796107f"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.397100 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.397475 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a2644986-db7d-4368-956b-65dc26585e3b","Type":"ContainerStarted","Data":"5aeda4e25b175fe0bac6bbc74c2e893dba7eea9418ca4bb73e023da4e4049158"} Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.617482 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.640866 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.698916 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.707185 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.733130 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.733962 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.788332 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.788417 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6wft\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.788466 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.792718 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.792773 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.792828 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.792861 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.792938 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.793034 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.800487 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.906511 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.906881 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6wft\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.906918 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.907045 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.907073 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.907102 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.907403 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.908318 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.908371 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.908424 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.908506 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.908890 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.914611 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.915342 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.919773 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.920455 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.922481 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.923119 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6wft\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:02 crc kubenswrapper[4888]: I1124 01:39:02.943591 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.033154 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.235686 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.278541 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.280469 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.282409 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.296068 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.309177 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.330494 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.409155 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7dfb67ff5b-2cglk"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.411328 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.428908 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.428963 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.428989 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.429052 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqqc4\" (UniqueName: \"kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.429088 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.429124 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.429148 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.429959 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dfb67ff5b-2cglk"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.452762 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.466181 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5a08f210-f6e8-46f3-9796-8f4236a53dd1","Type":"ContainerStarted","Data":"771d4c22575b2759b132e5bc8b131264287d306b771f92e954601821641b246d"} Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.466245 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5a08f210-f6e8-46f3-9796-8f4236a53dd1","Type":"ContainerStarted","Data":"3984e0d61234c60582ab7315fbd4ccec3a8e01cd098c77906edbe922f2245c15"} Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.472723 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a2644986-db7d-4368-956b-65dc26585e3b","Type":"ContainerStarted","Data":"6a101e7eaf4c1290c55350dc782a85a740b62ca60e217e9ff1490c5f801ac728"} Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.472760 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a2644986-db7d-4368-956b-65dc26585e3b","Type":"ContainerStarted","Data":"e28ed1e02d478ca3d51851690f89e04ad3f7a768e7ee5adc34e07ac788b5325f"} Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.505159 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.59601198 podStartE2EDuration="4.505134357s" podCreationTimestamp="2025-11-24 01:38:59 +0000 UTC" firstStartedPulling="2025-11-24 01:39:01.50313797 +0000 UTC m=+4444.085822014" lastFinishedPulling="2025-11-24 01:39:02.412260347 +0000 UTC m=+4444.994944391" observedRunningTime="2025-11-24 01:39:03.492848244 +0000 UTC m=+4446.075532288" watchObservedRunningTime="2025-11-24 01:39:03.505134357 +0000 UTC m=+4446.087818401" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.528216 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.505630686 podStartE2EDuration="4.528195892s" podCreationTimestamp="2025-11-24 01:38:59 +0000 UTC" firstStartedPulling="2025-11-24 01:39:01.363356575 +0000 UTC m=+4443.946040619" lastFinishedPulling="2025-11-24 01:39:02.385921781 +0000 UTC m=+4444.968605825" observedRunningTime="2025-11-24 01:39:03.513359107 +0000 UTC m=+4446.096043151" watchObservedRunningTime="2025-11-24 01:39:03.528195892 +0000 UTC m=+4446.110879936" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.533471 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-combined-ca-bundle\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.533543 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.534272 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.533579 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.534327 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-config-data\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535087 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535174 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqqc4\" (UniqueName: \"kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535199 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-tls-certs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535250 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8tkf\" (UniqueName: \"kubernetes.io/projected/78ef5328-d894-4921-8464-62bdaf1c4208-kube-api-access-j8tkf\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535287 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535311 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535335 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-scripts\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535362 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-secret-key\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535388 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ef5328-d894-4921-8464-62bdaf1c4208-logs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.535719 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.542018 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.544585 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.546320 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.568728 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqqc4\" (UniqueName: \"kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.573433 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle\") pod \"horizon-776b89d8d8-zqnfp\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637094 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-combined-ca-bundle\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637145 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-config-data\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-tls-certs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637300 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8tkf\" (UniqueName: \"kubernetes.io/projected/78ef5328-d894-4921-8464-62bdaf1c4208-kube-api-access-j8tkf\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637390 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-scripts\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637434 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-secret-key\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.637483 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ef5328-d894-4921-8464-62bdaf1c4208-logs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.643978 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-scripts\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.644450 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.645495 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78ef5328-d894-4921-8464-62bdaf1c4208-config-data\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.646208 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ef5328-d894-4921-8464-62bdaf1c4208-logs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.650438 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-tls-certs\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.661537 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-horizon-secret-key\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.663832 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ef5328-d894-4921-8464-62bdaf1c4208-combined-ca-bundle\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.684059 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8tkf\" (UniqueName: \"kubernetes.io/projected/78ef5328-d894-4921-8464-62bdaf1c4208-kube-api-access-j8tkf\") pod \"horizon-7dfb67ff5b-2cglk\" (UID: \"78ef5328-d894-4921-8464-62bdaf1c4208\") " pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.765585 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:03 crc kubenswrapper[4888]: I1124 01:39:03.900370 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.298397 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d62516-a39b-4748-944c-dbf1a52e2980" path="/var/lib/kubelet/pods/f2d62516-a39b-4748-944c-dbf1a52e2980/volumes" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.371469 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4j9db" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.477797 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts\") pod \"3a4de536-bef6-4f2b-ac46-64a43f970422\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.477935 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtczv\" (UniqueName: \"kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv\") pod \"3a4de536-bef6-4f2b-ac46-64a43f970422\" (UID: \"3a4de536-bef6-4f2b-ac46-64a43f970422\") " Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.478752 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a4de536-bef6-4f2b-ac46-64a43f970422" (UID: "3a4de536-bef6-4f2b-ac46-64a43f970422"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.484900 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv" (OuterVolumeSpecName: "kube-api-access-mtczv") pod "3a4de536-bef6-4f2b-ac46-64a43f970422" (UID: "3a4de536-bef6-4f2b-ac46-64a43f970422"). InnerVolumeSpecName "kube-api-access-mtczv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.497664 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.502654 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4j9db" event={"ID":"3a4de536-bef6-4f2b-ac46-64a43f970422","Type":"ContainerDied","Data":"635eec9e1ac134afdd1c158481f43d6088488286b26edbdac275c5b56723b1cc"} Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.502688 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="635eec9e1ac134afdd1c158481f43d6088488286b26edbdac275c5b56723b1cc" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.502738 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4j9db" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.506561 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-129e-account-create-8fkkk" event={"ID":"76da0164-f26f-445f-bc36-401a45149229","Type":"ContainerDied","Data":"69f077c2bf3ade7c2db8f0b59ea47b9e73f4d16b13ce4494fa0c476fd796107f"} Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.506601 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f077c2bf3ade7c2db8f0b59ea47b9e73f4d16b13ce4494fa0c476fd796107f" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.506802 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-129e-account-create-8fkkk" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.509972 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerStarted","Data":"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2"} Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.511639 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerStarted","Data":"249db86fa381e37fab4edd54258ae25932c21ffbeaff87d30c3e3d86ac32b544"} Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.530333 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:39:04 crc kubenswrapper[4888]: W1124 01:39:04.563247 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5e39f0_8f83_4d1a_93d9_ed8d4e713ce0.slice/crio-06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3 WatchSource:0}: Error finding container 06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3: Status 404 returned error can't find the container with id 06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3 Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.581059 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts\") pod \"76da0164-f26f-445f-bc36-401a45149229\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.581181 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7dvj\" (UniqueName: \"kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj\") pod \"76da0164-f26f-445f-bc36-401a45149229\" (UID: \"76da0164-f26f-445f-bc36-401a45149229\") " Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.581650 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76da0164-f26f-445f-bc36-401a45149229" (UID: "76da0164-f26f-445f-bc36-401a45149229"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.582954 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a4de536-bef6-4f2b-ac46-64a43f970422-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.582981 4888 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76da0164-f26f-445f-bc36-401a45149229-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.583009 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtczv\" (UniqueName: \"kubernetes.io/projected/3a4de536-bef6-4f2b-ac46-64a43f970422-kube-api-access-mtczv\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.606981 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj" (OuterVolumeSpecName: "kube-api-access-x7dvj") pod "76da0164-f26f-445f-bc36-401a45149229" (UID: "76da0164-f26f-445f-bc36-401a45149229"). InnerVolumeSpecName "kube-api-access-x7dvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.685156 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7dvj\" (UniqueName: \"kubernetes.io/projected/76da0164-f26f-445f-bc36-401a45149229-kube-api-access-x7dvj\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:04 crc kubenswrapper[4888]: I1124 01:39:04.755234 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dfb67ff5b-2cglk"] Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.135060 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.210317 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.543872 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerStarted","Data":"2c301951c3fe86dad2544d6d9190a7f7a47038c3eab2e53e4b893c40620ac2da"} Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.547081 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dfb67ff5b-2cglk" event={"ID":"78ef5328-d894-4921-8464-62bdaf1c4208","Type":"ContainerStarted","Data":"22b64af955d09cfb725fdc3df72ab7169c7ab08b078a4fc98d21d96d64a4910b"} Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.550375 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerStarted","Data":"06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3"} Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.564897 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerStarted","Data":"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6"} Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.564993 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-httpd" containerID="cri-o://7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" gracePeriod=30 Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.565187 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-log" containerID="cri-o://6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" gracePeriod=30 Nov 24 01:39:05 crc kubenswrapper[4888]: I1124 01:39:05.595547 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.595524894 podStartE2EDuration="5.595524894s" podCreationTimestamp="2025-11-24 01:39:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:05.588745324 +0000 UTC m=+4448.171429378" watchObservedRunningTime="2025-11-24 01:39:05.595524894 +0000 UTC m=+4448.178208948" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.084553 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-xg7z9"] Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.086393 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4de536-bef6-4f2b-ac46-64a43f970422" containerName="mariadb-database-create" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.086413 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4de536-bef6-4f2b-ac46-64a43f970422" containerName="mariadb-database-create" Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.086437 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76da0164-f26f-445f-bc36-401a45149229" containerName="mariadb-account-create" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.086443 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="76da0164-f26f-445f-bc36-401a45149229" containerName="mariadb-account-create" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.086966 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="76da0164-f26f-445f-bc36-401a45149229" containerName="mariadb-account-create" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.086997 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4de536-bef6-4f2b-ac46-64a43f970422" containerName="mariadb-database-create" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.088697 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.095315 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xg7z9"] Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.099972 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.100052 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-psdlk" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.223685 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.224052 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.224076 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.224130 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffhqm\" (UniqueName: \"kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.327226 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.327431 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.327454 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.327544 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffhqm\" (UniqueName: \"kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.338995 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.341553 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.352340 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.354986 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.359191 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffhqm\" (UniqueName: \"kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm\") pod \"manila-db-sync-xg7z9\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.431855 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.431999 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432091 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n74pf\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432168 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432431 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432509 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432698 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432795 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.432865 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run\") pod \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\" (UID: \"cd1dcf8e-55f1-45d1-bea9-3c685285b62f\") " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.436077 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs" (OuterVolumeSpecName: "logs") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.436242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.445351 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.459700 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf" (OuterVolumeSpecName: "kube-api-access-n74pf") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "kube-api-access-n74pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.477941 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.536930 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.536961 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.536974 4888 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.536985 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n74pf\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-kube-api-access-n74pf\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.564498 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph" (OuterVolumeSpecName: "ceph") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.564696 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts" (OuterVolumeSpecName: "scripts") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.565298 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.569842 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data" (OuterVolumeSpecName: "config-data") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.573916 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.617270 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cd1dcf8e-55f1-45d1-bea9-3c685285b62f" (UID: "cd1dcf8e-55f1-45d1-bea9-3c685285b62f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.640020 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerStarted","Data":"bf8b4be8e8478373d988c7d27c8a35989d90c657be2dcf44648ec4daae00c63f"} Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.640223 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-log" containerID="cri-o://2c301951c3fe86dad2544d6d9190a7f7a47038c3eab2e53e4b893c40620ac2da" gracePeriod=30 Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.640715 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-httpd" containerID="cri-o://bf8b4be8e8478373d988c7d27c8a35989d90c657be2dcf44648ec4daae00c63f" gracePeriod=30 Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642234 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642291 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642307 4888 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642321 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642338 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.642348 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cd1dcf8e-55f1-45d1-bea9-3c685285b62f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655519 4888 generic.go:334] "Generic (PLEG): container finished" podID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerID="7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" exitCode=143 Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655550 4888 generic.go:334] "Generic (PLEG): container finished" podID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerID="6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" exitCode=143 Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655572 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerDied","Data":"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6"} Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655599 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerDied","Data":"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2"} Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655608 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd1dcf8e-55f1-45d1-bea9-3c685285b62f","Type":"ContainerDied","Data":"78d5139d8bfaf12f30193b84469beeeb8fa7704ebae007c3655facb4a748ac76"} Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655626 4888 scope.go:117] "RemoveContainer" containerID="7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.655757 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.680688 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.680668758 podStartE2EDuration="4.680668758s" podCreationTimestamp="2025-11-24 01:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:06.668517409 +0000 UTC m=+4449.251201453" watchObservedRunningTime="2025-11-24 01:39:06.680668758 +0000 UTC m=+4449.263352792" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.710014 4888 scope.go:117] "RemoveContainer" containerID="6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.723540 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.735940 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.774985 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.775763 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-httpd" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.775783 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-httpd" Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.775805 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-log" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.775825 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-log" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.776060 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-log" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.776072 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" containerName="glance-httpd" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.777637 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.780338 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.781502 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.804458 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.816249 4888 scope.go:117] "RemoveContainer" containerID="7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.817749 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6\": container with ID starting with 7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6 not found: ID does not exist" containerID="7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.818662 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6"} err="failed to get container status \"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6\": rpc error: code = NotFound desc = could not find container \"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6\": container with ID starting with 7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6 not found: ID does not exist" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.818689 4888 scope.go:117] "RemoveContainer" containerID="6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" Nov 24 01:39:06 crc kubenswrapper[4888]: E1124 01:39:06.821776 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2\": container with ID starting with 6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2 not found: ID does not exist" containerID="6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.821830 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2"} err="failed to get container status \"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2\": rpc error: code = NotFound desc = could not find container \"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2\": container with ID starting with 6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2 not found: ID does not exist" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.821858 4888 scope.go:117] "RemoveContainer" containerID="7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.828171 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6"} err="failed to get container status \"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6\": rpc error: code = NotFound desc = could not find container \"7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6\": container with ID starting with 7e73b089d29a57a8ba21ce6093e488fdd8acb9d937f4d939a32bf548bf606db6 not found: ID does not exist" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.828212 4888 scope.go:117] "RemoveContainer" containerID="6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.828713 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2"} err="failed to get container status \"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2\": rpc error: code = NotFound desc = could not find container \"6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2\": container with ID starting with 6a2a0f1c984dbb7a69a27aba02f3ad77d45326d9eae24a959ea331cea8e62fb2 not found: ID does not exist" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.860529 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863229 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2qhj\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-kube-api-access-j2qhj\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863349 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863573 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863616 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863641 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863762 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-ceph\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863826 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-logs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.863941 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.967455 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2qhj\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-kube-api-access-j2qhj\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.967582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.967608 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.968198 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.968739 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969191 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969268 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969344 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-ceph\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969445 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-logs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969558 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.969608 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.971397 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5296f4e7-f819-43bb-9566-ee12ba6d51f0-logs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.979702 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.980380 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-ceph\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.982117 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.983606 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.984071 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5296f4e7-f819-43bb-9566-ee12ba6d51f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:06 crc kubenswrapper[4888]: I1124 01:39:06.992249 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2qhj\" (UniqueName: \"kubernetes.io/projected/5296f4e7-f819-43bb-9566-ee12ba6d51f0-kube-api-access-j2qhj\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.009672 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5296f4e7-f819-43bb-9566-ee12ba6d51f0\") " pod="openstack/glance-default-external-api-0" Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.098940 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.367230 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xg7z9"] Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.697785 4888 generic.go:334] "Generic (PLEG): container finished" podID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerID="bf8b4be8e8478373d988c7d27c8a35989d90c657be2dcf44648ec4daae00c63f" exitCode=0 Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.697838 4888 generic.go:334] "Generic (PLEG): container finished" podID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerID="2c301951c3fe86dad2544d6d9190a7f7a47038c3eab2e53e4b893c40620ac2da" exitCode=143 Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.697933 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerDied","Data":"bf8b4be8e8478373d988c7d27c8a35989d90c657be2dcf44648ec4daae00c63f"} Nov 24 01:39:07 crc kubenswrapper[4888]: I1124 01:39:07.697968 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerDied","Data":"2c301951c3fe86dad2544d6d9190a7f7a47038c3eab2e53e4b893c40620ac2da"} Nov 24 01:39:07 crc kubenswrapper[4888]: W1124 01:39:07.885061 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04e08e0b_9532_4dee_92ff_10324d806d08.slice/crio-c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5 WatchSource:0}: Error finding container c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5: Status 404 returned error can't find the container with id c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5 Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.067929 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.145545 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.145628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.145718 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.145783 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6wft\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.145942 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.146193 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.146605 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.146698 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.146718 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.146786 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph\") pod \"30db5b75-bec6-417b-88dc-01eaaceb4623\" (UID: \"30db5b75-bec6-417b-88dc-01eaaceb4623\") " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.147387 4888 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.151065 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs" (OuterVolumeSpecName: "logs") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.154841 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft" (OuterVolumeSpecName: "kube-api-access-n6wft") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "kube-api-access-n6wft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.154918 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.158097 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph" (OuterVolumeSpecName: "ceph") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.163629 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts" (OuterVolumeSpecName: "scripts") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.210879 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259791 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259842 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259854 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6wft\" (UniqueName: \"kubernetes.io/projected/30db5b75-bec6-417b-88dc-01eaaceb4623-kube-api-access-n6wft\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259865 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259873 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.259881 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30db5b75-bec6-417b-88dc-01eaaceb4623-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.262961 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data" (OuterVolumeSpecName: "config-data") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.284532 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1dcf8e-55f1-45d1-bea9-3c685285b62f" path="/var/lib/kubelet/pods/cd1dcf8e-55f1-45d1-bea9-3c685285b62f/volumes" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.291320 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "30db5b75-bec6-417b-88dc-01eaaceb4623" (UID: "30db5b75-bec6-417b-88dc-01eaaceb4623"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.293710 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.365802 4888 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.365872 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30db5b75-bec6-417b-88dc-01eaaceb4623-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.365886 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.522048 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.728780 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5296f4e7-f819-43bb-9566-ee12ba6d51f0","Type":"ContainerStarted","Data":"2285e8277a3845f15389f353464d28bec706412e7eeaec329eaf299870b615f0"} Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.732839 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xg7z9" event={"ID":"04e08e0b-9532-4dee-92ff-10324d806d08","Type":"ContainerStarted","Data":"c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5"} Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.738858 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"30db5b75-bec6-417b-88dc-01eaaceb4623","Type":"ContainerDied","Data":"249db86fa381e37fab4edd54258ae25932c21ffbeaff87d30c3e3d86ac32b544"} Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.738908 4888 scope.go:117] "RemoveContainer" containerID="bf8b4be8e8478373d988c7d27c8a35989d90c657be2dcf44648ec4daae00c63f" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.739063 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.777981 4888 scope.go:117] "RemoveContainer" containerID="2c301951c3fe86dad2544d6d9190a7f7a47038c3eab2e53e4b893c40620ac2da" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.800895 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.813752 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.822778 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:08 crc kubenswrapper[4888]: E1124 01:39:08.823570 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-log" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.823590 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-log" Nov 24 01:39:08 crc kubenswrapper[4888]: E1124 01:39:08.823607 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-httpd" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.823613 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-httpd" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.823884 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-httpd" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.823898 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" containerName="glance-log" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.825870 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.833204 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.833487 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.834724 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.878473 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.878670 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.878739 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.878773 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.878856 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.879132 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.879223 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdq87\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-kube-api-access-kdq87\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.879468 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.879584 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.984568 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985050 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985187 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985266 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985300 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985333 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985374 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985434 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.985478 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdq87\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-kube-api-access-kdq87\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.986291 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.986312 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.986429 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2ae1990-d967-4fe3-80cf-9520e5b83286-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.991709 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.992333 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.992545 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.992790 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:08 crc kubenswrapper[4888]: I1124 01:39:08.994773 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ae1990-d967-4fe3-80cf-9520e5b83286-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:09 crc kubenswrapper[4888]: I1124 01:39:09.001241 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdq87\" (UniqueName: \"kubernetes.io/projected/a2ae1990-d967-4fe3-80cf-9520e5b83286-kube-api-access-kdq87\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:09 crc kubenswrapper[4888]: I1124 01:39:09.037140 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2ae1990-d967-4fe3-80cf-9520e5b83286\") " pod="openstack/glance-default-internal-api-0" Nov 24 01:39:09 crc kubenswrapper[4888]: I1124 01:39:09.153266 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:09 crc kubenswrapper[4888]: I1124 01:39:09.760230 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5296f4e7-f819-43bb-9566-ee12ba6d51f0","Type":"ContainerStarted","Data":"6010be6db174335436d72dec6138233912c0ac9bce52e59c323937c5f8702daf"} Nov 24 01:39:09 crc kubenswrapper[4888]: I1124 01:39:09.797850 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 01:39:10 crc kubenswrapper[4888]: I1124 01:39:10.284091 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30db5b75-bec6-417b-88dc-01eaaceb4623" path="/var/lib/kubelet/pods/30db5b75-bec6-417b-88dc-01eaaceb4623/volumes" Nov 24 01:39:10 crc kubenswrapper[4888]: I1124 01:39:10.347311 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 24 01:39:10 crc kubenswrapper[4888]: I1124 01:39:10.442414 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 24 01:39:15 crc kubenswrapper[4888]: I1124 01:39:15.244867 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:39:15 crc kubenswrapper[4888]: E1124 01:39:15.245993 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:39:16 crc kubenswrapper[4888]: W1124 01:39:16.980630 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2ae1990_d967_4fe3_80cf_9520e5b83286.slice/crio-8da7688bbe98bcf12e5934670ce9b9a3245d06cccf24af2cf5e72c860c9a94ac WatchSource:0}: Error finding container 8da7688bbe98bcf12e5934670ce9b9a3245d06cccf24af2cf5e72c860c9a94ac: Status 404 returned error can't find the container with id 8da7688bbe98bcf12e5934670ce9b9a3245d06cccf24af2cf5e72c860c9a94ac Nov 24 01:39:17 crc kubenswrapper[4888]: I1124 01:39:17.865973 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2ae1990-d967-4fe3-80cf-9520e5b83286","Type":"ContainerStarted","Data":"8da7688bbe98bcf12e5934670ce9b9a3245d06cccf24af2cf5e72c860c9a94ac"} Nov 24 01:39:18 crc kubenswrapper[4888]: I1124 01:39:18.879131 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerStarted","Data":"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a"} Nov 24 01:39:18 crc kubenswrapper[4888]: I1124 01:39:18.881849 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerStarted","Data":"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637"} Nov 24 01:39:18 crc kubenswrapper[4888]: I1124 01:39:18.883963 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dfb67ff5b-2cglk" event={"ID":"78ef5328-d894-4921-8464-62bdaf1c4208","Type":"ContainerStarted","Data":"6896da971729f36dbf881ef1b94b4c3d396e0bfba5a52c051e9f0635ecd54983"} Nov 24 01:39:18 crc kubenswrapper[4888]: I1124 01:39:18.885471 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerStarted","Data":"af15b819f8a3240dea553f709b65bef99f9954d0fae76056e04ecb24ffb06fe4"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.900226 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerStarted","Data":"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.900359 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d8d9c897-jxm7k" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon-log" containerID="cri-o://439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" gracePeriod=30 Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.900416 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5d8d9c897-jxm7k" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon" containerID="cri-o://51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" gracePeriod=30 Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.903530 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dfb67ff5b-2cglk" event={"ID":"78ef5328-d894-4921-8464-62bdaf1c4208","Type":"ContainerStarted","Data":"f07582731ff37c00b041794f3546174a18fa7eedacf9a14469a21c40b6be3d72"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.912279 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2ae1990-d967-4fe3-80cf-9520e5b83286","Type":"ContainerStarted","Data":"457b411fe24e49c507b9781b002e4ff4472d68ead9ae97e0f54e21ebc8ec2f17"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.912324 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2ae1990-d967-4fe3-80cf-9520e5b83286","Type":"ContainerStarted","Data":"3f678066761aa9e7e4dc8a8939ddfb06add1a15857bc9ca8e19c259d96e00743"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.914862 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5296f4e7-f819-43bb-9566-ee12ba6d51f0","Type":"ContainerStarted","Data":"4fb45c58d93f1e1e05cc2fe55d096c08e8d3dbd032f0b0b75317f1d9c5a3133f"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.919966 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerStarted","Data":"c241f05f63597ea7ddc5c5f59643721b337b65e85987cbf0c54851c79c296917"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.922340 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerStarted","Data":"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.922485 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66644f99b9-l6svl" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon-log" containerID="cri-o://cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" gracePeriod=30 Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.922625 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66644f99b9-l6svl" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon" containerID="cri-o://3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" gracePeriod=30 Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.927883 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xg7z9" event={"ID":"04e08e0b-9532-4dee-92ff-10324d806d08","Type":"ContainerStarted","Data":"d3c87ca81b5b759dd447e6d765e75144f5d70321af7010a070c54364d7844742"} Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.931882 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5d8d9c897-jxm7k" podStartSLOduration=3.4545722039999998 podStartE2EDuration="19.931860058s" podCreationTimestamp="2025-11-24 01:39:00 +0000 UTC" firstStartedPulling="2025-11-24 01:39:01.879659058 +0000 UTC m=+4444.462343102" lastFinishedPulling="2025-11-24 01:39:18.356946912 +0000 UTC m=+4460.939630956" observedRunningTime="2025-11-24 01:39:19.927366403 +0000 UTC m=+4462.510050477" watchObservedRunningTime="2025-11-24 01:39:19.931860058 +0000 UTC m=+4462.514544102" Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.958278 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=13.958255896 podStartE2EDuration="13.958255896s" podCreationTimestamp="2025-11-24 01:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:19.947009472 +0000 UTC m=+4462.529693556" watchObservedRunningTime="2025-11-24 01:39:19.958255896 +0000 UTC m=+4462.540939950" Nov 24 01:39:19 crc kubenswrapper[4888]: I1124 01:39:19.972309 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-776b89d8d8-zqnfp" podStartSLOduration=3.186008599 podStartE2EDuration="16.972291878s" podCreationTimestamp="2025-11-24 01:39:03 +0000 UTC" firstStartedPulling="2025-11-24 01:39:04.569915852 +0000 UTC m=+4447.152599896" lastFinishedPulling="2025-11-24 01:39:18.356199131 +0000 UTC m=+4460.938883175" observedRunningTime="2025-11-24 01:39:19.96877218 +0000 UTC m=+4462.551456234" watchObservedRunningTime="2025-11-24 01:39:19.972291878 +0000 UTC m=+4462.554975922" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.006717 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.006697399 podStartE2EDuration="12.006697399s" podCreationTimestamp="2025-11-24 01:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:19.991637118 +0000 UTC m=+4462.574321162" watchObservedRunningTime="2025-11-24 01:39:20.006697399 +0000 UTC m=+4462.589381443" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.019049 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66644f99b9-l6svl" podStartSLOduration=3.108695412 podStartE2EDuration="20.019024243s" podCreationTimestamp="2025-11-24 01:39:00 +0000 UTC" firstStartedPulling="2025-11-24 01:39:01.445638063 +0000 UTC m=+4444.028322107" lastFinishedPulling="2025-11-24 01:39:18.355966884 +0000 UTC m=+4460.938650938" observedRunningTime="2025-11-24 01:39:20.017498271 +0000 UTC m=+4462.600182325" watchObservedRunningTime="2025-11-24 01:39:20.019024243 +0000 UTC m=+4462.601708297" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.048753 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7dfb67ff5b-2cglk" podStartSLOduration=3.533823745 podStartE2EDuration="17.048734873s" podCreationTimestamp="2025-11-24 01:39:03 +0000 UTC" firstStartedPulling="2025-11-24 01:39:04.776540274 +0000 UTC m=+4447.359224308" lastFinishedPulling="2025-11-24 01:39:18.291451382 +0000 UTC m=+4460.874135436" observedRunningTime="2025-11-24 01:39:20.044940837 +0000 UTC m=+4462.627624881" watchObservedRunningTime="2025-11-24 01:39:20.048734873 +0000 UTC m=+4462.631418927" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.074823 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-xg7z9" podStartSLOduration=3.6152218769999998 podStartE2EDuration="14.074793481s" podCreationTimestamp="2025-11-24 01:39:06 +0000 UTC" firstStartedPulling="2025-11-24 01:39:07.896062661 +0000 UTC m=+4450.478746695" lastFinishedPulling="2025-11-24 01:39:18.355634245 +0000 UTC m=+4460.938318299" observedRunningTime="2025-11-24 01:39:20.064420112 +0000 UTC m=+4462.647104166" watchObservedRunningTime="2025-11-24 01:39:20.074793481 +0000 UTC m=+4462.657477525" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.779435 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:20 crc kubenswrapper[4888]: I1124 01:39:20.987267 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:23 crc kubenswrapper[4888]: I1124 01:39:23.645257 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:23 crc kubenswrapper[4888]: I1124 01:39:23.645709 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:23 crc kubenswrapper[4888]: I1124 01:39:23.766717 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:23 crc kubenswrapper[4888]: I1124 01:39:23.766762 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:27 crc kubenswrapper[4888]: I1124 01:39:27.099104 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 01:39:27 crc kubenswrapper[4888]: I1124 01:39:27.099374 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 01:39:27 crc kubenswrapper[4888]: I1124 01:39:27.135544 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 01:39:27 crc kubenswrapper[4888]: I1124 01:39:27.145017 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 01:39:28 crc kubenswrapper[4888]: I1124 01:39:28.022585 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 01:39:28 crc kubenswrapper[4888]: I1124 01:39:28.022904 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 01:39:28 crc kubenswrapper[4888]: I1124 01:39:28.253954 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:39:28 crc kubenswrapper[4888]: E1124 01:39:28.254204 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:39:29 crc kubenswrapper[4888]: I1124 01:39:29.155115 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:29 crc kubenswrapper[4888]: I1124 01:39:29.155257 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:29 crc kubenswrapper[4888]: I1124 01:39:29.210607 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:29 crc kubenswrapper[4888]: I1124 01:39:29.224174 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.046348 4888 generic.go:334] "Generic (PLEG): container finished" podID="04e08e0b-9532-4dee-92ff-10324d806d08" containerID="d3c87ca81b5b759dd447e6d765e75144f5d70321af7010a070c54364d7844742" exitCode=0 Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.046931 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xg7z9" event={"ID":"04e08e0b-9532-4dee-92ff-10324d806d08","Type":"ContainerDied","Data":"d3c87ca81b5b759dd447e6d765e75144f5d70321af7010a070c54364d7844742"} Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.047018 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.047039 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.929827 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.930187 4888 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 01:39:30 crc kubenswrapper[4888]: I1124 01:39:30.972886 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.086956 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.087868 4888 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.117087 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.386518 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.507126 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffhqm\" (UniqueName: \"kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm\") pod \"04e08e0b-9532-4dee-92ff-10324d806d08\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.507954 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle\") pod \"04e08e0b-9532-4dee-92ff-10324d806d08\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.508030 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data\") pod \"04e08e0b-9532-4dee-92ff-10324d806d08\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.508155 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data\") pod \"04e08e0b-9532-4dee-92ff-10324d806d08\" (UID: \"04e08e0b-9532-4dee-92ff-10324d806d08\") " Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.516284 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm" (OuterVolumeSpecName: "kube-api-access-ffhqm") pod "04e08e0b-9532-4dee-92ff-10324d806d08" (UID: "04e08e0b-9532-4dee-92ff-10324d806d08"). InnerVolumeSpecName "kube-api-access-ffhqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.521153 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "04e08e0b-9532-4dee-92ff-10324d806d08" (UID: "04e08e0b-9532-4dee-92ff-10324d806d08"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.522487 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data" (OuterVolumeSpecName: "config-data") pod "04e08e0b-9532-4dee-92ff-10324d806d08" (UID: "04e08e0b-9532-4dee-92ff-10324d806d08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.550536 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04e08e0b-9532-4dee-92ff-10324d806d08" (UID: "04e08e0b-9532-4dee-92ff-10324d806d08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.610711 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.610750 4888 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.610761 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e08e0b-9532-4dee-92ff-10324d806d08-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:32 crc kubenswrapper[4888]: I1124 01:39:32.610773 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffhqm\" (UniqueName: \"kubernetes.io/projected/04e08e0b-9532-4dee-92ff-10324d806d08-kube-api-access-ffhqm\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.098135 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xg7z9" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.098180 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xg7z9" event={"ID":"04e08e0b-9532-4dee-92ff-10324d806d08","Type":"ContainerDied","Data":"c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5"} Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.098207 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c20ce580003d8b2999ba7a6986b039f5a396404ee6a26c18fa77dd52f89705f5" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.647476 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.68:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.68:8443: connect: connection refused" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.727776 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:39:33 crc kubenswrapper[4888]: E1124 01:39:33.728250 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e08e0b-9532-4dee-92ff-10324d806d08" containerName="manila-db-sync" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.728265 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e08e0b-9532-4dee-92ff-10324d806d08" containerName="manila-db-sync" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.728483 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e08e0b-9532-4dee-92ff-10324d806d08" containerName="manila-db-sync" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.732017 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.737711 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.738237 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-psdlk" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.742913 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.747073 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.758412 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.793911 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7dfb67ff5b-2cglk" podUID="78ef5328-d894-4921-8464-62bdaf1c4208" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.884489 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.884970 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.885225 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.885336 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgkt7\" (UniqueName: \"kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.885485 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.885850 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.912919 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.929015 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.935651 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.943850 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.953786 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-k75jx"] Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.955729 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:33 crc kubenswrapper[4888]: I1124 01:39:33.972533 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-k75jx"] Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003293 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003358 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003413 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003492 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003522 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgkt7\" (UniqueName: \"kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.003566 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.004894 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.038933 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.040995 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.045519 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.056520 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105544 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105603 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105635 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-config\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105673 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105708 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105756 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105776 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105917 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105945 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.105961 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.106007 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.106031 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp82l\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.106089 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.106107 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md848\" (UniqueName: \"kubernetes.io/projected/9d8de233-4a00-482c-b35f-170370015b11-kube-api-access-md848\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.106140 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.151093 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.151177 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.151604 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.153483 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.155549 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgkt7\" (UniqueName: \"kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7\") pod \"manila-scheduler-0\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208632 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208677 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208727 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208752 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208779 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208797 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208837 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5k7f\" (UniqueName: \"kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208854 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208872 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208888 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208903 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208948 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.208968 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp82l\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209024 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209040 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md848\" (UniqueName: \"kubernetes.io/projected/9d8de233-4a00-482c-b35f-170370015b11-kube-api-access-md848\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209073 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209095 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209116 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209134 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209155 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-config\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209185 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209216 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.209290 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.210161 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.210194 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.210856 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.211170 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.211255 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.211381 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.211627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8de233-4a00-482c-b35f-170370015b11-config\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.213392 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.214552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.217131 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.231514 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md848\" (UniqueName: \"kubernetes.io/projected/9d8de233-4a00-482c-b35f-170370015b11-kube-api-access-md848\") pod \"dnsmasq-dns-5767ddb7c-k75jx\" (UID: \"9d8de233-4a00-482c-b35f-170370015b11\") " pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.231541 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp82l\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.244878 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.263618 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.266610 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.314600 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315368 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5k7f\" (UniqueName: \"kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315419 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315554 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315622 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315643 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315687 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.315704 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.316670 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.317235 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.323427 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.324182 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.324544 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.328215 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.339625 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5k7f\" (UniqueName: \"kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f\") pod \"manila-api-0\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " pod="openstack/manila-api-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.373747 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:39:34 crc kubenswrapper[4888]: I1124 01:39:34.380850 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:35 crc kubenswrapper[4888]: I1124 01:39:35.082395 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:39:35 crc kubenswrapper[4888]: I1124 01:39:35.190398 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-k75jx"] Nov 24 01:39:35 crc kubenswrapper[4888]: I1124 01:39:35.394416 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:39:35 crc kubenswrapper[4888]: I1124 01:39:35.554980 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:35 crc kubenswrapper[4888]: W1124 01:39:35.564628 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6657e73_b395_405f_919f_ce09b545c74c.slice/crio-facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f WatchSource:0}: Error finding container facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f: Status 404 returned error can't find the container with id facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.133048 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerStarted","Data":"f45f2a9d50e312ac5dcd379208eac2d8d5f73d18d9eee99783500981013cb661"} Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.134803 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerStarted","Data":"f50539478ff08dc8df8a461a56c2e90ae0da10d157feb90d0e04f5dbcd342759"} Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.134866 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerStarted","Data":"facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f"} Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.137731 4888 generic.go:334] "Generic (PLEG): container finished" podID="9d8de233-4a00-482c-b35f-170370015b11" containerID="b2d54c0e9d9c0570630c2735c961919f64b0d4c81286cb61d854b4cd88572815" exitCode=0 Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.137782 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" event={"ID":"9d8de233-4a00-482c-b35f-170370015b11","Type":"ContainerDied","Data":"b2d54c0e9d9c0570630c2735c961919f64b0d4c81286cb61d854b4cd88572815"} Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.137844 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" event={"ID":"9d8de233-4a00-482c-b35f-170370015b11","Type":"ContainerStarted","Data":"0213b749ed3857e27d6f04d3b3cb086db0dfcf5a12e23e9a815528371818db97"} Nov 24 01:39:36 crc kubenswrapper[4888]: I1124 01:39:36.139268 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerStarted","Data":"e94a494de2c94cf8d36efc545999606b96975456076f82931cbda5fcb22c7ab3"} Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.253150 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerStarted","Data":"2aa1f9522922d5ae085428107ad51ef4234301638302bd5b3e2319f2a0813556"} Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.253785 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.280361 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" event={"ID":"9d8de233-4a00-482c-b35f-170370015b11","Type":"ContainerStarted","Data":"18c749890687bc9919da270991879648173176911e56ca1f0d567292e81b5ee4"} Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.284075 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.290306 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.290286436 podStartE2EDuration="4.290286436s" podCreationTimestamp="2025-11-24 01:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:37.284442553 +0000 UTC m=+4479.867126597" watchObservedRunningTime="2025-11-24 01:39:37.290286436 +0000 UTC m=+4479.872970490" Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.308290 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerStarted","Data":"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed"} Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.341300 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" podStartSLOduration=4.341280891 podStartE2EDuration="4.341280891s" podCreationTimestamp="2025-11-24 01:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:37.314445291 +0000 UTC m=+4479.897129335" watchObservedRunningTime="2025-11-24 01:39:37.341280891 +0000 UTC m=+4479.923964925" Nov 24 01:39:37 crc kubenswrapper[4888]: I1124 01:39:37.544671 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:38 crc kubenswrapper[4888]: I1124 01:39:38.326339 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerStarted","Data":"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d"} Nov 24 01:39:38 crc kubenswrapper[4888]: I1124 01:39:38.360270 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.41533856 podStartE2EDuration="5.360252236s" podCreationTimestamp="2025-11-24 01:39:33 +0000 UTC" firstStartedPulling="2025-11-24 01:39:35.448994029 +0000 UTC m=+4478.031678073" lastFinishedPulling="2025-11-24 01:39:36.393907705 +0000 UTC m=+4478.976591749" observedRunningTime="2025-11-24 01:39:38.355241237 +0000 UTC m=+4480.937925281" watchObservedRunningTime="2025-11-24 01:39:38.360252236 +0000 UTC m=+4480.942936290" Nov 24 01:39:39 crc kubenswrapper[4888]: I1124 01:39:39.334247 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api-log" containerID="cri-o://f50539478ff08dc8df8a461a56c2e90ae0da10d157feb90d0e04f5dbcd342759" gracePeriod=30 Nov 24 01:39:39 crc kubenswrapper[4888]: I1124 01:39:39.334500 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api" containerID="cri-o://2aa1f9522922d5ae085428107ad51ef4234301638302bd5b3e2319f2a0813556" gracePeriod=30 Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344024 4888 generic.go:334] "Generic (PLEG): container finished" podID="f6657e73-b395-405f-919f-ce09b545c74c" containerID="2aa1f9522922d5ae085428107ad51ef4234301638302bd5b3e2319f2a0813556" exitCode=0 Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344498 4888 generic.go:334] "Generic (PLEG): container finished" podID="f6657e73-b395-405f-919f-ce09b545c74c" containerID="f50539478ff08dc8df8a461a56c2e90ae0da10d157feb90d0e04f5dbcd342759" exitCode=143 Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344069 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerDied","Data":"2aa1f9522922d5ae085428107ad51ef4234301638302bd5b3e2319f2a0813556"} Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344534 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerDied","Data":"f50539478ff08dc8df8a461a56c2e90ae0da10d157feb90d0e04f5dbcd342759"} Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344551 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6657e73-b395-405f-919f-ce09b545c74c","Type":"ContainerDied","Data":"facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f"} Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.344560 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="facd4d3a88dd61f15451397217e12646dd708ef365421281d7f6888e5d431b8f" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.432213 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.594748 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5k7f\" (UniqueName: \"kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.594792 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.594821 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.594954 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595035 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595175 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595253 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595318 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts\") pod \"f6657e73-b395-405f-919f-ce09b545c74c\" (UID: \"f6657e73-b395-405f-919f-ce09b545c74c\") " Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595383 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs" (OuterVolumeSpecName: "logs") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595844 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6657e73-b395-405f-919f-ce09b545c74c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.595861 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6657e73-b395-405f-919f-ce09b545c74c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.600855 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f" (OuterVolumeSpecName: "kube-api-access-r5k7f") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "kube-api-access-r5k7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.602029 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts" (OuterVolumeSpecName: "scripts") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.602572 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.630519 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.659599 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data" (OuterVolumeSpecName: "config-data") pod "f6657e73-b395-405f-919f-ce09b545c74c" (UID: "f6657e73-b395-405f-919f-ce09b545c74c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.698455 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.698491 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.698500 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.698509 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5k7f\" (UniqueName: \"kubernetes.io/projected/f6657e73-b395-405f-919f-ce09b545c74c-kube-api-access-r5k7f\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:40 crc kubenswrapper[4888]: I1124 01:39:40.698520 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6657e73-b395-405f-919f-ce09b545c74c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.354618 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.395487 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.403934 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.426104 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:41 crc kubenswrapper[4888]: E1124 01:39:41.426547 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.426560 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api" Nov 24 01:39:41 crc kubenswrapper[4888]: E1124 01:39:41.426578 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api-log" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.426585 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api-log" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.426834 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api-log" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.426851 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6657e73-b395-405f-919f-ce09b545c74c" containerName="manila-api" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.428082 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.431199 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.431236 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.431400 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.448041 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.620499 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data-custom\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.620577 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bcac458-0aa6-4a6c-8586-666803573def-logs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.620648 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lm7w\" (UniqueName: \"kubernetes.io/projected/8bcac458-0aa6-4a6c-8586-666803573def-kube-api-access-5lm7w\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.620926 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bcac458-0aa6-4a6c-8586-666803573def-etc-machine-id\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.620994 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.621152 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-scripts\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.621210 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.621307 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.621368 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-public-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.723949 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bcac458-0aa6-4a6c-8586-666803573def-etc-machine-id\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724014 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724044 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bcac458-0aa6-4a6c-8586-666803573def-etc-machine-id\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724077 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-scripts\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724114 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724164 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724198 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-public-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data-custom\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724296 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bcac458-0aa6-4a6c-8586-666803573def-logs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.724354 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lm7w\" (UniqueName: \"kubernetes.io/projected/8bcac458-0aa6-4a6c-8586-666803573def-kube-api-access-5lm7w\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.725069 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bcac458-0aa6-4a6c-8586-666803573def-logs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.729534 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-scripts\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.730491 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data-custom\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.730984 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.733109 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.733408 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-config-data\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.733770 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcac458-0aa6-4a6c-8586-666803573def-public-tls-certs\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.740193 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lm7w\" (UniqueName: \"kubernetes.io/projected/8bcac458-0aa6-4a6c-8586-666803573def-kube-api-access-5lm7w\") pod \"manila-api-0\" (UID: \"8bcac458-0aa6-4a6c-8586-666803573def\") " pod="openstack/manila-api-0" Nov 24 01:39:41 crc kubenswrapper[4888]: I1124 01:39:41.772771 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.263425 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6657e73-b395-405f-919f-ce09b545c74c" path="/var/lib/kubelet/pods/f6657e73-b395-405f-919f-ce09b545c74c/volumes" Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.375545 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.376895 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-central-agent" containerID="cri-o://587ed132ea34f3ebadc888b2b3c0e909483c9272d0497faeded2b6ebb3fc128b" gracePeriod=30 Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.377018 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-notification-agent" containerID="cri-o://6aba00bfee962dd890d93a5867f963b103ae0c8dbc9c95abcc626cdcd747548e" gracePeriod=30 Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.377153 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="sg-core" containerID="cri-o://02d1227a7ec42fa3f8790c2eb85236a26e280b2afc33ed698fcc045e07cb6fe7" gracePeriod=30 Nov 24 01:39:42 crc kubenswrapper[4888]: I1124 01:39:42.376959 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="proxy-httpd" containerID="cri-o://d7c28ede11c7103214633efc6e8d69d4b1af684a72e184c5f75ccbcb8349256d" gracePeriod=30 Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.245657 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:39:43 crc kubenswrapper[4888]: E1124 01:39:43.246254 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387432 4888 generic.go:334] "Generic (PLEG): container finished" podID="ca359a50-4d08-4b99-99eb-091732d34061" containerID="d7c28ede11c7103214633efc6e8d69d4b1af684a72e184c5f75ccbcb8349256d" exitCode=0 Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387705 4888 generic.go:334] "Generic (PLEG): container finished" podID="ca359a50-4d08-4b99-99eb-091732d34061" containerID="02d1227a7ec42fa3f8790c2eb85236a26e280b2afc33ed698fcc045e07cb6fe7" exitCode=2 Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387713 4888 generic.go:334] "Generic (PLEG): container finished" podID="ca359a50-4d08-4b99-99eb-091732d34061" containerID="587ed132ea34f3ebadc888b2b3c0e909483c9272d0497faeded2b6ebb3fc128b" exitCode=0 Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387527 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerDied","Data":"d7c28ede11c7103214633efc6e8d69d4b1af684a72e184c5f75ccbcb8349256d"} Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387748 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerDied","Data":"02d1227a7ec42fa3f8790c2eb85236a26e280b2afc33ed698fcc045e07cb6fe7"} Nov 24 01:39:43 crc kubenswrapper[4888]: I1124 01:39:43.387762 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerDied","Data":"587ed132ea34f3ebadc888b2b3c0e909483c9272d0497faeded2b6ebb3fc128b"} Nov 24 01:39:44 crc kubenswrapper[4888]: I1124 01:39:44.195274 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 01:39:44 crc kubenswrapper[4888]: I1124 01:39:44.317058 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5767ddb7c-k75jx" Nov 24 01:39:44 crc kubenswrapper[4888]: I1124 01:39:44.374015 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 01:39:44 crc kubenswrapper[4888]: I1124 01:39:44.413678 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 01:39:44 crc kubenswrapper[4888]: I1124 01:39:44.413964 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="dnsmasq-dns" containerID="cri-o://665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720" gracePeriod=10 Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.260488 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416205 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416349 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416409 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg6rm\" (UniqueName: \"kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416454 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416471 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416494 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.416554 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam\") pod \"377504e9-345b-4365-a0b4-3922c341e4d9\" (UID: \"377504e9-345b-4365-a0b4-3922c341e4d9\") " Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.420695 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm" (OuterVolumeSpecName: "kube-api-access-dg6rm") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "kube-api-access-dg6rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.448157 4888 generic.go:334] "Generic (PLEG): container finished" podID="377504e9-345b-4365-a0b4-3922c341e4d9" containerID="665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720" exitCode=0 Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.448220 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.448229 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" event={"ID":"377504e9-345b-4365-a0b4-3922c341e4d9","Type":"ContainerDied","Data":"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720"} Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.448262 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-m5kg8" event={"ID":"377504e9-345b-4365-a0b4-3922c341e4d9","Type":"ContainerDied","Data":"76e5183791a01f810c373f3ce6f270b772fece3df443563cb92f7d33c51aca19"} Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.448280 4888 scope.go:117] "RemoveContainer" containerID="665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.457377 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerStarted","Data":"515c450f600bede2918ee2a441c7b01a06083340cfbedabf222a57829ee5da1d"} Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.464482 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8bcac458-0aa6-4a6c-8586-666803573def","Type":"ContainerStarted","Data":"8738f1199966a79b67bc2bc3b9e78116ce64f759b1b3e77ba678cffd8bc3ba8f"} Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.464526 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8bcac458-0aa6-4a6c-8586-666803573def","Type":"ContainerStarted","Data":"995a50fe11f90c7dba24e6faee6b93d60e153cf62b3d431b07dd8210d18b2599"} Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.507486 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.512035 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.513575 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.520910 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.520988 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.521010 4888 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.521020 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.521028 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg6rm\" (UniqueName: \"kubernetes.io/projected/377504e9-345b-4365-a0b4-3922c341e4d9-kube-api-access-dg6rm\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.521459 4888 scope.go:117] "RemoveContainer" containerID="ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.525156 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.533981 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config" (OuterVolumeSpecName: "config") pod "377504e9-345b-4365-a0b4-3922c341e4d9" (UID: "377504e9-345b-4365-a0b4-3922c341e4d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.547749 4888 scope.go:117] "RemoveContainer" containerID="665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720" Nov 24 01:39:45 crc kubenswrapper[4888]: E1124 01:39:45.548189 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720\": container with ID starting with 665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720 not found: ID does not exist" containerID="665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.548246 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720"} err="failed to get container status \"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720\": rpc error: code = NotFound desc = could not find container \"665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720\": container with ID starting with 665a51935a7b1aa19bde474940651b2e08ea33b04209b9c595865eb6cd1f0720 not found: ID does not exist" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.548282 4888 scope.go:117] "RemoveContainer" containerID="ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a" Nov 24 01:39:45 crc kubenswrapper[4888]: E1124 01:39:45.548671 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a\": container with ID starting with ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a not found: ID does not exist" containerID="ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.548705 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a"} err="failed to get container status \"ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a\": rpc error: code = NotFound desc = could not find container \"ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a\": container with ID starting with ce0094fe66635a67c670834d2333d84383e5e85c7b9ce367fd5f3301fe72f35a not found: ID does not exist" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.622838 4888 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.623220 4888 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-config\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.623232 4888 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/377504e9-345b-4365-a0b4-3922c341e4d9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.830794 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 01:39:45 crc kubenswrapper[4888]: I1124 01:39:45.843172 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-m5kg8"] Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.177302 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.227830 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.293994 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" path="/var/lib/kubelet/pods/377504e9-345b-4365-a0b4-3922c341e4d9/volumes" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.505901 4888 generic.go:334] "Generic (PLEG): container finished" podID="ca359a50-4d08-4b99-99eb-091732d34061" containerID="6aba00bfee962dd890d93a5867f963b103ae0c8dbc9c95abcc626cdcd747548e" exitCode=0 Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.505987 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerDied","Data":"6aba00bfee962dd890d93a5867f963b103ae0c8dbc9c95abcc626cdcd747548e"} Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.508270 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerStarted","Data":"d8ec94040fb07cbf993713b53167ae2733ef9c9f0b35bd0f23df9a8f032b30e2"} Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.517973 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8bcac458-0aa6-4a6c-8586-666803573def","Type":"ContainerStarted","Data":"2df9b77896dbe73a8ea0a834dfc0bf956d594a0b160bf4603a2bd3fcae1627e6"} Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.518185 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.540651 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=5.012628107 podStartE2EDuration="13.540630562s" podCreationTimestamp="2025-11-24 01:39:33 +0000 UTC" firstStartedPulling="2025-11-24 01:39:35.154459161 +0000 UTC m=+4477.737143195" lastFinishedPulling="2025-11-24 01:39:43.682461606 +0000 UTC m=+4486.265145650" observedRunningTime="2025-11-24 01:39:46.534072329 +0000 UTC m=+4489.116756383" watchObservedRunningTime="2025-11-24 01:39:46.540630562 +0000 UTC m=+4489.123314606" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.581773 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.581747411 podStartE2EDuration="5.581747411s" podCreationTimestamp="2025-11-24 01:39:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:39:46.553108411 +0000 UTC m=+4489.135792455" watchObservedRunningTime="2025-11-24 01:39:46.581747411 +0000 UTC m=+4489.164431495" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.628295 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.755193 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.756318 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.756470 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.756566 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.756631 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.757121 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.757179 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h7j5\" (UniqueName: \"kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.757206 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs\") pod \"ca359a50-4d08-4b99-99eb-091732d34061\" (UID: \"ca359a50-4d08-4b99-99eb-091732d34061\") " Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.762630 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.764314 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.794390 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts" (OuterVolumeSpecName: "scripts") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.795795 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5" (OuterVolumeSpecName: "kube-api-access-8h7j5") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "kube-api-access-8h7j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.808959 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.838398 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862350 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862381 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862391 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862403 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca359a50-4d08-4b99-99eb-091732d34061-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862413 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h7j5\" (UniqueName: \"kubernetes.io/projected/ca359a50-4d08-4b99-99eb-091732d34061-kube-api-access-8h7j5\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.862422 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.864926 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.937578 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data" (OuterVolumeSpecName: "config-data") pod "ca359a50-4d08-4b99-99eb-091732d34061" (UID: "ca359a50-4d08-4b99-99eb-091732d34061"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.965939 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:46 crc kubenswrapper[4888]: I1124 01:39:46.965978 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca359a50-4d08-4b99-99eb-091732d34061-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.528925 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca359a50-4d08-4b99-99eb-091732d34061","Type":"ContainerDied","Data":"c02f9cfdeb7bbf6a4b4ffb6eb9662e2922098ccedb060ecff4541896b33d7a17"} Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.529276 4888 scope.go:117] "RemoveContainer" containerID="d7c28ede11c7103214633efc6e8d69d4b1af684a72e184c5f75ccbcb8349256d" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.529107 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.585074 4888 scope.go:117] "RemoveContainer" containerID="02d1227a7ec42fa3f8790c2eb85236a26e280b2afc33ed698fcc045e07cb6fe7" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.608921 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.620829 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.632988 4888 scope.go:117] "RemoveContainer" containerID="6aba00bfee962dd890d93a5867f963b103ae0c8dbc9c95abcc626cdcd747548e" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635102 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635537 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="proxy-httpd" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635551 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="proxy-httpd" Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635572 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="init" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635578 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="init" Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635592 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-notification-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635598 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-notification-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635631 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="sg-core" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635637 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="sg-core" Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635646 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-central-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635651 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-central-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: E1124 01:39:47.635664 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="dnsmasq-dns" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635669 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="dnsmasq-dns" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635901 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="377504e9-345b-4365-a0b4-3922c341e4d9" containerName="dnsmasq-dns" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635919 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-notification-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635930 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="ceilometer-central-agent" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635947 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="proxy-httpd" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.635963 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca359a50-4d08-4b99-99eb-091732d34061" containerName="sg-core" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.638270 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.640774 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.641090 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.641293 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.665186 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.687697 4888 scope.go:117] "RemoveContainer" containerID="587ed132ea34f3ebadc888b2b3c0e909483c9272d0497faeded2b6ebb3fc128b" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789129 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789191 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2br6\" (UniqueName: \"kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789238 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789293 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789671 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789801 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789926 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.789988 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.891835 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.891900 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892008 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892045 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2br6\" (UniqueName: \"kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892089 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892143 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892280 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892309 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.892638 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.894522 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.897718 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.900147 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.900281 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.909908 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.910128 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.916463 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2br6\" (UniqueName: \"kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6\") pod \"ceilometer-0\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.964183 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:47 crc kubenswrapper[4888]: I1124 01:39:47.982094 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.128831 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7dfb67ff5b-2cglk" Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.192505 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.260575 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca359a50-4d08-4b99-99eb-091732d34061" path="/var/lib/kubelet/pods/ca359a50-4d08-4b99-99eb-091732d34061/volumes" Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.510377 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:48 crc kubenswrapper[4888]: W1124 01:39:48.520503 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d083447_c1f1_4ccc_8adb_1e10422a1a6e.slice/crio-8d070a880dc211d260194babc78fa6fac3c06fb0407cc629780288412b836b80 WatchSource:0}: Error finding container 8d070a880dc211d260194babc78fa6fac3c06fb0407cc629780288412b836b80: Status 404 returned error can't find the container with id 8d070a880dc211d260194babc78fa6fac3c06fb0407cc629780288412b836b80 Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.547274 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerStarted","Data":"8d070a880dc211d260194babc78fa6fac3c06fb0407cc629780288412b836b80"} Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.552378 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon-log" containerID="cri-o://af15b819f8a3240dea553f709b65bef99f9954d0fae76056e04ecb24ffb06fe4" gracePeriod=30 Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.552433 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" containerID="cri-o://c241f05f63597ea7ddc5c5f59643721b337b65e85987cbf0c54851c79c296917" gracePeriod=30 Nov 24 01:39:48 crc kubenswrapper[4888]: I1124 01:39:48.636720 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:49 crc kubenswrapper[4888]: I1124 01:39:49.566612 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerStarted","Data":"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.420379 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.490068 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.570698 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key\") pod \"99da32f4-1f0b-4d1b-9456-63c4df848427\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.570966 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts\") pod \"99da32f4-1f0b-4d1b-9456-63c4df848427\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.571027 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs\") pod \"99da32f4-1f0b-4d1b-9456-63c4df848427\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.571097 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfqwm\" (UniqueName: \"kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm\") pod \"99da32f4-1f0b-4d1b-9456-63c4df848427\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.571118 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data\") pod \"99da32f4-1f0b-4d1b-9456-63c4df848427\" (UID: \"99da32f4-1f0b-4d1b-9456-63c4df848427\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.573507 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs" (OuterVolumeSpecName: "logs") pod "99da32f4-1f0b-4d1b-9456-63c4df848427" (UID: "99da32f4-1f0b-4d1b-9456-63c4df848427"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.574849 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "99da32f4-1f0b-4d1b-9456-63c4df848427" (UID: "99da32f4-1f0b-4d1b-9456-63c4df848427"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.576987 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm" (OuterVolumeSpecName: "kube-api-access-hfqwm") pod "99da32f4-1f0b-4d1b-9456-63c4df848427" (UID: "99da32f4-1f0b-4d1b-9456-63c4df848427"). InnerVolumeSpecName "kube-api-access-hfqwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583172 4888 generic.go:334] "Generic (PLEG): container finished" podID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerID="3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" exitCode=137 Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583201 4888 generic.go:334] "Generic (PLEG): container finished" podID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerID="cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" exitCode=137 Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583237 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerDied","Data":"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583263 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerDied","Data":"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583273 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66644f99b9-l6svl" event={"ID":"99da32f4-1f0b-4d1b-9456-63c4df848427","Type":"ContainerDied","Data":"011fc40162721804cb934c5b76405a6e8a41759158e63facabbfd35a6b4fa0ef"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583287 4888 scope.go:117] "RemoveContainer" containerID="3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.583393 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66644f99b9-l6svl" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588114 4888 generic.go:334] "Generic (PLEG): container finished" podID="cce8d408-043a-42be-8d65-a9a96c78670a" containerID="51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" exitCode=137 Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588177 4888 generic.go:334] "Generic (PLEG): container finished" podID="cce8d408-043a-42be-8d65-a9a96c78670a" containerID="439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" exitCode=137 Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588256 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerDied","Data":"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588285 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerDied","Data":"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588296 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d8d9c897-jxm7k" event={"ID":"cce8d408-043a-42be-8d65-a9a96c78670a","Type":"ContainerDied","Data":"d22ad00a23030c923813d4eb06f0cb13f25b769aa77a079eac685fa6c70f575f"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.588350 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d8d9c897-jxm7k" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.595159 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerStarted","Data":"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.595197 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerStarted","Data":"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6"} Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.610325 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data" (OuterVolumeSpecName: "config-data") pod "99da32f4-1f0b-4d1b-9456-63c4df848427" (UID: "99da32f4-1f0b-4d1b-9456-63c4df848427"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.613037 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts" (OuterVolumeSpecName: "scripts") pod "99da32f4-1f0b-4d1b-9456-63c4df848427" (UID: "99da32f4-1f0b-4d1b-9456-63c4df848427"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.672725 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr5f5\" (UniqueName: \"kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5\") pod \"cce8d408-043a-42be-8d65-a9a96c78670a\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.674123 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts\") pod \"cce8d408-043a-42be-8d65-a9a96c78670a\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.674184 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data\") pod \"cce8d408-043a-42be-8d65-a9a96c78670a\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.674233 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key\") pod \"cce8d408-043a-42be-8d65-a9a96c78670a\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.674346 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs\") pod \"cce8d408-043a-42be-8d65-a9a96c78670a\" (UID: \"cce8d408-043a-42be-8d65-a9a96c78670a\") " Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675270 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs" (OuterVolumeSpecName: "logs") pod "cce8d408-043a-42be-8d65-a9a96c78670a" (UID: "cce8d408-043a-42be-8d65-a9a96c78670a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675287 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675334 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99da32f4-1f0b-4d1b-9456-63c4df848427-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675348 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfqwm\" (UniqueName: \"kubernetes.io/projected/99da32f4-1f0b-4d1b-9456-63c4df848427-kube-api-access-hfqwm\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675359 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99da32f4-1f0b-4d1b-9456-63c4df848427-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.675398 4888 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99da32f4-1f0b-4d1b-9456-63c4df848427-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.676051 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5" (OuterVolumeSpecName: "kube-api-access-zr5f5") pod "cce8d408-043a-42be-8d65-a9a96c78670a" (UID: "cce8d408-043a-42be-8d65-a9a96c78670a"). InnerVolumeSpecName "kube-api-access-zr5f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.679548 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cce8d408-043a-42be-8d65-a9a96c78670a" (UID: "cce8d408-043a-42be-8d65-a9a96c78670a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.709937 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data" (OuterVolumeSpecName: "config-data") pod "cce8d408-043a-42be-8d65-a9a96c78670a" (UID: "cce8d408-043a-42be-8d65-a9a96c78670a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.712787 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts" (OuterVolumeSpecName: "scripts") pod "cce8d408-043a-42be-8d65-a9a96c78670a" (UID: "cce8d408-043a-42be-8d65-a9a96c78670a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.769612 4888 scope.go:117] "RemoveContainer" containerID="cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.777268 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.777309 4888 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cce8d408-043a-42be-8d65-a9a96c78670a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.777323 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cce8d408-043a-42be-8d65-a9a96c78670a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.777335 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr5f5\" (UniqueName: \"kubernetes.io/projected/cce8d408-043a-42be-8d65-a9a96c78670a-kube-api-access-zr5f5\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.777347 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cce8d408-043a-42be-8d65-a9a96c78670a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.793635 4888 scope.go:117] "RemoveContainer" containerID="3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" Nov 24 01:39:50 crc kubenswrapper[4888]: E1124 01:39:50.794177 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113\": container with ID starting with 3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113 not found: ID does not exist" containerID="3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.794216 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113"} err="failed to get container status \"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113\": rpc error: code = NotFound desc = could not find container \"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113\": container with ID starting with 3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113 not found: ID does not exist" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.794244 4888 scope.go:117] "RemoveContainer" containerID="cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" Nov 24 01:39:50 crc kubenswrapper[4888]: E1124 01:39:50.794798 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a\": container with ID starting with cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a not found: ID does not exist" containerID="cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.794871 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a"} err="failed to get container status \"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a\": rpc error: code = NotFound desc = could not find container \"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a\": container with ID starting with cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a not found: ID does not exist" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.794905 4888 scope.go:117] "RemoveContainer" containerID="3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.795244 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113"} err="failed to get container status \"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113\": rpc error: code = NotFound desc = could not find container \"3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113\": container with ID starting with 3301a90b44f0b51da91ebd2a282a7acb4afab8b851f221c9c8844eff42d66113 not found: ID does not exist" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.795271 4888 scope.go:117] "RemoveContainer" containerID="cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.795689 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a"} err="failed to get container status \"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a\": rpc error: code = NotFound desc = could not find container \"cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a\": container with ID starting with cd4eb7a5e215e02ffb11f846c131fea509f49152d9f7d70e1cb21a135dd16d1a not found: ID does not exist" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.795736 4888 scope.go:117] "RemoveContainer" containerID="51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.937730 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.955459 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66644f99b9-l6svl"] Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.970273 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.974261 4888 scope.go:117] "RemoveContainer" containerID="439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" Nov 24 01:39:50 crc kubenswrapper[4888]: I1124 01:39:50.984388 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5d8d9c897-jxm7k"] Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.075357 4888 scope.go:117] "RemoveContainer" containerID="51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" Nov 24 01:39:51 crc kubenswrapper[4888]: E1124 01:39:51.075736 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff\": container with ID starting with 51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff not found: ID does not exist" containerID="51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.075764 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff"} err="failed to get container status \"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff\": rpc error: code = NotFound desc = could not find container \"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff\": container with ID starting with 51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff not found: ID does not exist" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.075791 4888 scope.go:117] "RemoveContainer" containerID="439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" Nov 24 01:39:51 crc kubenswrapper[4888]: E1124 01:39:51.076269 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637\": container with ID starting with 439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637 not found: ID does not exist" containerID="439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.076298 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637"} err="failed to get container status \"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637\": rpc error: code = NotFound desc = could not find container \"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637\": container with ID starting with 439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637 not found: ID does not exist" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.076313 4888 scope.go:117] "RemoveContainer" containerID="51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.076658 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff"} err="failed to get container status \"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff\": rpc error: code = NotFound desc = could not find container \"51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff\": container with ID starting with 51e8d9b19ee52ba9dde6b63b73d475b432c18b71548e4f91a2a84fac31506cff not found: ID does not exist" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.076683 4888 scope.go:117] "RemoveContainer" containerID="439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637" Nov 24 01:39:51 crc kubenswrapper[4888]: I1124 01:39:51.076874 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637"} err="failed to get container status \"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637\": rpc error: code = NotFound desc = could not find container \"439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637\": container with ID starting with 439d5a3a69d2dbd01ab39be8bc0b24df540613f3c68640f1618fde2238161637 not found: ID does not exist" Nov 24 01:39:52 crc kubenswrapper[4888]: I1124 01:39:52.260440 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" path="/var/lib/kubelet/pods/99da32f4-1f0b-4d1b-9456-63c4df848427/volumes" Nov 24 01:39:52 crc kubenswrapper[4888]: I1124 01:39:52.261612 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" path="/var/lib/kubelet/pods/cce8d408-043a-42be-8d65-a9a96c78670a/volumes" Nov 24 01:39:52 crc kubenswrapper[4888]: I1124 01:39:52.618462 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerID="c241f05f63597ea7ddc5c5f59643721b337b65e85987cbf0c54851c79c296917" exitCode=0 Nov 24 01:39:52 crc kubenswrapper[4888]: I1124 01:39:52.618544 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerDied","Data":"c241f05f63597ea7ddc5c5f59643721b337b65e85987cbf0c54851c79c296917"} Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.632386 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerStarted","Data":"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17"} Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.633172 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-central-agent" containerID="cri-o://4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda" gracePeriod=30 Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.634018 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.634402 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="proxy-httpd" containerID="cri-o://73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17" gracePeriod=30 Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.634484 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="sg-core" containerID="cri-o://b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9" gracePeriod=30 Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.634530 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-notification-agent" containerID="cri-o://351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6" gracePeriod=30 Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.646750 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.68:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.68:8443: connect: connection refused" Nov 24 01:39:53 crc kubenswrapper[4888]: I1124 01:39:53.678736 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.260714165 podStartE2EDuration="6.678715009s" podCreationTimestamp="2025-11-24 01:39:47 +0000 UTC" firstStartedPulling="2025-11-24 01:39:48.522647991 +0000 UTC m=+4491.105332045" lastFinishedPulling="2025-11-24 01:39:51.940648845 +0000 UTC m=+4494.523332889" observedRunningTime="2025-11-24 01:39:53.667073194 +0000 UTC m=+4496.249757268" watchObservedRunningTime="2025-11-24 01:39:53.678715009 +0000 UTC m=+4496.261399063" Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.267595 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.651890 4888 generic.go:334] "Generic (PLEG): container finished" podID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerID="73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17" exitCode=0 Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.651947 4888 generic.go:334] "Generic (PLEG): container finished" podID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerID="b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9" exitCode=2 Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.651957 4888 generic.go:334] "Generic (PLEG): container finished" podID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerID="351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6" exitCode=0 Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.651975 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerDied","Data":"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17"} Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.652042 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerDied","Data":"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9"} Nov 24 01:39:54 crc kubenswrapper[4888]: I1124 01:39:54.652073 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerDied","Data":"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6"} Nov 24 01:39:55 crc kubenswrapper[4888]: I1124 01:39:55.246060 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:39:55 crc kubenswrapper[4888]: E1124 01:39:55.246681 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:39:55 crc kubenswrapper[4888]: I1124 01:39:55.823197 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 01:39:55 crc kubenswrapper[4888]: I1124 01:39:55.890189 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:39:56 crc kubenswrapper[4888]: I1124 01:39:56.670635 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="manila-scheduler" containerID="cri-o://0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed" gracePeriod=30 Nov 24 01:39:56 crc kubenswrapper[4888]: I1124 01:39:56.670792 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="probe" containerID="cri-o://be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d" gracePeriod=30 Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.188576 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336417 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336563 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336613 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336742 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336767 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336808 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336876 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2br6\" (UniqueName: \"kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.336911 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd\") pod \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\" (UID: \"0d083447-c1f1-4ccc-8adb-1e10422a1a6e\") " Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.337116 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.337507 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.337651 4888 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.337669 4888 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.343321 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts" (OuterVolumeSpecName: "scripts") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.343837 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6" (OuterVolumeSpecName: "kube-api-access-f2br6") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "kube-api-access-f2br6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.372670 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.418856 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.439528 4888 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.439553 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.439565 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2br6\" (UniqueName: \"kubernetes.io/projected/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-kube-api-access-f2br6\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.439578 4888 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.440793 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.460605 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data" (OuterVolumeSpecName: "config-data") pod "0d083447-c1f1-4ccc-8adb-1e10422a1a6e" (UID: "0d083447-c1f1-4ccc-8adb-1e10422a1a6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.542054 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.542096 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d083447-c1f1-4ccc-8adb-1e10422a1a6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.688660 4888 generic.go:334] "Generic (PLEG): container finished" podID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerID="4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda" exitCode=0 Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.688717 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerDied","Data":"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda"} Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.688748 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.688780 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d083447-c1f1-4ccc-8adb-1e10422a1a6e","Type":"ContainerDied","Data":"8d070a880dc211d260194babc78fa6fac3c06fb0407cc629780288412b836b80"} Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.688799 4888 scope.go:117] "RemoveContainer" containerID="73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.693332 4888 generic.go:334] "Generic (PLEG): container finished" podID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerID="be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d" exitCode=0 Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.693366 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerDied","Data":"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d"} Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.735207 4888 scope.go:117] "RemoveContainer" containerID="b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.745485 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.762842 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.785226 4888 scope.go:117] "RemoveContainer" containerID="351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.792711 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793272 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="sg-core" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793291 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="sg-core" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793312 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-central-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793322 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-central-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793347 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793358 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793446 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793458 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793474 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793480 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793491 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-notification-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793498 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-notification-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793521 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="proxy-httpd" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793527 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="proxy-httpd" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.793549 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793555 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793753 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-central-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793763 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793777 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="proxy-httpd" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793791 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="ceilometer-notification-agent" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793965 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" containerName="sg-core" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.793980 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.794030 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="99da32f4-1f0b-4d1b-9456-63c4df848427" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.794047 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce8d408-043a-42be-8d65-a9a96c78670a" containerName="horizon-log" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.797696 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.802067 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.802851 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.802872 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.812625 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.870696 4888 scope.go:117] "RemoveContainer" containerID="4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.902993 4888 scope.go:117] "RemoveContainer" containerID="73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.905358 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17\": container with ID starting with 73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17 not found: ID does not exist" containerID="73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.905416 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17"} err="failed to get container status \"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17\": rpc error: code = NotFound desc = could not find container \"73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17\": container with ID starting with 73ac201d1b55d5ff4387c2987cda9b536780c598634975df561c788a2b628a17 not found: ID does not exist" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.905449 4888 scope.go:117] "RemoveContainer" containerID="b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.905765 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9\": container with ID starting with b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9 not found: ID does not exist" containerID="b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.905799 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9"} err="failed to get container status \"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9\": rpc error: code = NotFound desc = could not find container \"b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9\": container with ID starting with b2aec75e2c0b785f5c33bdca0b9640fcb1d41f3ef4519efc9ee8d6a957494ef9 not found: ID does not exist" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.905865 4888 scope.go:117] "RemoveContainer" containerID="351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.906499 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6\": container with ID starting with 351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6 not found: ID does not exist" containerID="351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.906527 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6"} err="failed to get container status \"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6\": rpc error: code = NotFound desc = could not find container \"351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6\": container with ID starting with 351d5be0e0df55b64c6497f4fb8a6bce8538aa73889ef4af9f99d66760b37fb6 not found: ID does not exist" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.906545 4888 scope.go:117] "RemoveContainer" containerID="4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda" Nov 24 01:39:57 crc kubenswrapper[4888]: E1124 01:39:57.906844 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda\": container with ID starting with 4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda not found: ID does not exist" containerID="4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.906917 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda"} err="failed to get container status \"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda\": rpc error: code = NotFound desc = could not find container \"4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda\": container with ID starting with 4e5dc0f245dd2122a5d57cb39512d898d491d5c52799205d5cf29ac9ed054cda not found: ID does not exist" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.956366 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-scripts\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.956626 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-run-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.956766 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.956904 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.957008 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.957120 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-config-data\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.957235 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5nvb\" (UniqueName: \"kubernetes.io/projected/9ea8705f-eabc-47ad-9f38-c77011ba6616-kube-api-access-x5nvb\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:57 crc kubenswrapper[4888]: I1124 01:39:57.957536 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-log-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.059699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-run-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.060031 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.060190 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.060419 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-run-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.060966 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.061111 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-config-data\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.061150 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5nvb\" (UniqueName: \"kubernetes.io/projected/9ea8705f-eabc-47ad-9f38-c77011ba6616-kube-api-access-x5nvb\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.061607 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-log-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.061899 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-scripts\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.062244 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ea8705f-eabc-47ad-9f38-c77011ba6616-log-httpd\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.065014 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-config-data\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.065143 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.065631 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.067350 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-scripts\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.070430 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ea8705f-eabc-47ad-9f38-c77011ba6616-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.079221 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5nvb\" (UniqueName: \"kubernetes.io/projected/9ea8705f-eabc-47ad-9f38-c77011ba6616-kube-api-access-x5nvb\") pod \"ceilometer-0\" (UID: \"9ea8705f-eabc-47ad-9f38-c77011ba6616\") " pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.159703 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.274728 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d083447-c1f1-4ccc-8adb-1e10422a1a6e" path="/var/lib/kubelet/pods/0d083447-c1f1-4ccc-8adb-1e10422a1a6e/volumes" Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.650749 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 01:39:58 crc kubenswrapper[4888]: I1124 01:39:58.710320 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ea8705f-eabc-47ad-9f38-c77011ba6616","Type":"ContainerStarted","Data":"8e91fab07cf1afc4bafd6d598d7f5d66e2f75c1e3e0b53f42ece6e3b458ba45a"} Nov 24 01:39:59 crc kubenswrapper[4888]: I1124 01:39:59.724340 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ea8705f-eabc-47ad-9f38-c77011ba6616","Type":"ContainerStarted","Data":"8ed02b38cdde85fd7afab8ffb02a6d704d27e84583a0a56a0fc34bb73ced0d5d"} Nov 24 01:39:59 crc kubenswrapper[4888]: I1124 01:39:59.725077 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ea8705f-eabc-47ad-9f38-c77011ba6616","Type":"ContainerStarted","Data":"c2f78129a74cd2e22449d173cbadaa9403392c4aefcf9bad360a6d7ebf5e98d8"} Nov 24 01:40:00 crc kubenswrapper[4888]: I1124 01:40:00.753445 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ea8705f-eabc-47ad-9f38-c77011ba6616","Type":"ContainerStarted","Data":"6cbd293b78a17d4e24055e4ce5e82c580b6d311ee58054bc0a56640939f7cfe0"} Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.485369 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.640532 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.640715 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgkt7\" (UniqueName: \"kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.641650 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.641730 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.641829 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.641893 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id\") pod \"e8383367-0f04-4b4a-b7a3-240db042af2c\" (UID: \"e8383367-0f04-4b4a-b7a3-240db042af2c\") " Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.642753 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.648116 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.648350 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7" (OuterVolumeSpecName: "kube-api-access-lgkt7") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "kube-api-access-lgkt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.649404 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts" (OuterVolumeSpecName: "scripts") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.711073 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.744894 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.744932 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.744947 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.744960 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e8383367-0f04-4b4a-b7a3-240db042af2c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.744971 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgkt7\" (UniqueName: \"kubernetes.io/projected/e8383367-0f04-4b4a-b7a3-240db042af2c-kube-api-access-lgkt7\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.767025 4888 generic.go:334] "Generic (PLEG): container finished" podID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerID="0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed" exitCode=0 Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.767070 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerDied","Data":"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed"} Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.767096 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e8383367-0f04-4b4a-b7a3-240db042af2c","Type":"ContainerDied","Data":"e94a494de2c94cf8d36efc545999606b96975456076f82931cbda5fcb22c7ab3"} Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.767112 4888 scope.go:117] "RemoveContainer" containerID="be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.767228 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.783042 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data" (OuterVolumeSpecName: "config-data") pod "e8383367-0f04-4b4a-b7a3-240db042af2c" (UID: "e8383367-0f04-4b4a-b7a3-240db042af2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.793754 4888 scope.go:117] "RemoveContainer" containerID="0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.818936 4888 scope.go:117] "RemoveContainer" containerID="be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d" Nov 24 01:40:01 crc kubenswrapper[4888]: E1124 01:40:01.822088 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d\": container with ID starting with be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d not found: ID does not exist" containerID="be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.822132 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d"} err="failed to get container status \"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d\": rpc error: code = NotFound desc = could not find container \"be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d\": container with ID starting with be231c2c1be21d682111bc3298386cab2ce4d0040bdb6e30935e84fa0327f03d not found: ID does not exist" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.822167 4888 scope.go:117] "RemoveContainer" containerID="0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed" Nov 24 01:40:01 crc kubenswrapper[4888]: E1124 01:40:01.822544 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed\": container with ID starting with 0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed not found: ID does not exist" containerID="0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.822575 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed"} err="failed to get container status \"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed\": rpc error: code = NotFound desc = could not find container \"0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed\": container with ID starting with 0324958fe6b53e9383b00e650f381d31593549d51efe071c1ad6a825ed8b83ed not found: ID does not exist" Nov 24 01:40:01 crc kubenswrapper[4888]: I1124 01:40:01.846917 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8383367-0f04-4b4a-b7a3-240db042af2c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.150594 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.168006 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.181851 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:40:02 crc kubenswrapper[4888]: E1124 01:40:02.182453 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="probe" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.182473 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="probe" Nov 24 01:40:02 crc kubenswrapper[4888]: E1124 01:40:02.182506 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="manila-scheduler" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.182514 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="manila-scheduler" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.182846 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="probe" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.182865 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" containerName="manila-scheduler" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.184447 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.187689 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.190072 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257113 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfhgf\" (UniqueName: \"kubernetes.io/projected/eff408f2-f760-40a0-9016-b149c398cf85-kube-api-access-sfhgf\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257279 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257324 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eff408f2-f760-40a0-9016-b149c398cf85-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257393 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257438 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.257482 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-scripts\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.260264 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8383367-0f04-4b4a-b7a3-240db042af2c" path="/var/lib/kubelet/pods/e8383367-0f04-4b4a-b7a3-240db042af2c/volumes" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.360713 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.360783 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eff408f2-f760-40a0-9016-b149c398cf85-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.360888 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.361014 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.361061 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-scripts\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.361296 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfhgf\" (UniqueName: \"kubernetes.io/projected/eff408f2-f760-40a0-9016-b149c398cf85-kube-api-access-sfhgf\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.364487 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eff408f2-f760-40a0-9016-b149c398cf85-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.370898 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-scripts\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.370938 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.372268 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.373837 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff408f2-f760-40a0-9016-b149c398cf85-config-data\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.382121 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfhgf\" (UniqueName: \"kubernetes.io/projected/eff408f2-f760-40a0-9016-b149c398cf85-kube-api-access-sfhgf\") pod \"manila-scheduler-0\" (UID: \"eff408f2-f760-40a0-9016-b149c398cf85\") " pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.518703 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.781380 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ea8705f-eabc-47ad-9f38-c77011ba6616","Type":"ContainerStarted","Data":"eb1d51e3c9bb6a4cc614340e2a73cb47b99ab909b731e2f92671e5517fe71db7"} Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.782953 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.815508 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.565220742 podStartE2EDuration="5.8154867s" podCreationTimestamp="2025-11-24 01:39:57 +0000 UTC" firstStartedPulling="2025-11-24 01:39:58.655517028 +0000 UTC m=+4501.238201072" lastFinishedPulling="2025-11-24 01:40:01.905782986 +0000 UTC m=+4504.488467030" observedRunningTime="2025-11-24 01:40:02.802102996 +0000 UTC m=+4505.384787040" watchObservedRunningTime="2025-11-24 01:40:02.8154867 +0000 UTC m=+4505.398170744" Nov 24 01:40:02 crc kubenswrapper[4888]: I1124 01:40:02.965216 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 01:40:03 crc kubenswrapper[4888]: I1124 01:40:03.153186 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 24 01:40:03 crc kubenswrapper[4888]: I1124 01:40:03.645891 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.68:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.68:8443: connect: connection refused" Nov 24 01:40:03 crc kubenswrapper[4888]: I1124 01:40:03.793564 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eff408f2-f760-40a0-9016-b149c398cf85","Type":"ContainerStarted","Data":"86bed317f520ee070a340f5ed144fa57d5f045a567e9c8857322acb272c659b8"} Nov 24 01:40:04 crc kubenswrapper[4888]: I1124 01:40:04.807494 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eff408f2-f760-40a0-9016-b149c398cf85","Type":"ContainerStarted","Data":"5043eaee51f9340107c8ab7baa445dad1f0c0b7a544cb738a51d64b288143e12"} Nov 24 01:40:04 crc kubenswrapper[4888]: I1124 01:40:04.808042 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eff408f2-f760-40a0-9016-b149c398cf85","Type":"ContainerStarted","Data":"6cbbb74215792147f6d8448164ba8f0eb8a65e283fd645d7a37c6e86dd2f2191"} Nov 24 01:40:04 crc kubenswrapper[4888]: I1124 01:40:04.846452 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.846432475 podStartE2EDuration="2.846432475s" podCreationTimestamp="2025-11-24 01:40:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:40:04.838226706 +0000 UTC m=+4507.420910750" watchObservedRunningTime="2025-11-24 01:40:04.846432475 +0000 UTC m=+4507.429116519" Nov 24 01:40:05 crc kubenswrapper[4888]: I1124 01:40:05.834239 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 01:40:05 crc kubenswrapper[4888]: I1124 01:40:05.918124 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:06 crc kubenswrapper[4888]: I1124 01:40:06.836048 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="manila-share" containerID="cri-o://515c450f600bede2918ee2a441c7b01a06083340cfbedabf222a57829ee5da1d" gracePeriod=30 Nov 24 01:40:06 crc kubenswrapper[4888]: I1124 01:40:06.836115 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="probe" containerID="cri-o://d8ec94040fb07cbf993713b53167ae2733ef9c9f0b35bd0f23df9a8f032b30e2" gracePeriod=30 Nov 24 01:40:07 crc kubenswrapper[4888]: I1124 01:40:07.850646 4888 generic.go:334] "Generic (PLEG): container finished" podID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerID="d8ec94040fb07cbf993713b53167ae2733ef9c9f0b35bd0f23df9a8f032b30e2" exitCode=0 Nov 24 01:40:07 crc kubenswrapper[4888]: I1124 01:40:07.851240 4888 generic.go:334] "Generic (PLEG): container finished" podID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerID="515c450f600bede2918ee2a441c7b01a06083340cfbedabf222a57829ee5da1d" exitCode=1 Nov 24 01:40:07 crc kubenswrapper[4888]: I1124 01:40:07.850713 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerDied","Data":"d8ec94040fb07cbf993713b53167ae2733ef9c9f0b35bd0f23df9a8f032b30e2"} Nov 24 01:40:07 crc kubenswrapper[4888]: I1124 01:40:07.851289 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerDied","Data":"515c450f600bede2918ee2a441c7b01a06083340cfbedabf222a57829ee5da1d"} Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.003744 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097081 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097268 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097262 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097322 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097399 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097450 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097510 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp82l\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097767 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.097793 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts\") pod \"65746f6a-e726-41e1-adb1-33d86a5ad858\" (UID: \"65746f6a-e726-41e1-adb1-33d86a5ad858\") " Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.098186 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.098447 4888 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.098477 4888 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65746f6a-e726-41e1-adb1-33d86a5ad858-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.111269 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph" (OuterVolumeSpecName: "ceph") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.111326 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l" (OuterVolumeSpecName: "kube-api-access-cp82l") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "kube-api-access-cp82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.111350 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts" (OuterVolumeSpecName: "scripts") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.111410 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.169904 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.200304 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.200343 4888 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.200355 4888 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.200369 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp82l\" (UniqueName: \"kubernetes.io/projected/65746f6a-e726-41e1-adb1-33d86a5ad858-kube-api-access-cp82l\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.200381 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.229653 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data" (OuterVolumeSpecName: "config-data") pod "65746f6a-e726-41e1-adb1-33d86a5ad858" (UID: "65746f6a-e726-41e1-adb1-33d86a5ad858"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.302178 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65746f6a-e726-41e1-adb1-33d86a5ad858-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.866080 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"65746f6a-e726-41e1-adb1-33d86a5ad858","Type":"ContainerDied","Data":"f45f2a9d50e312ac5dcd379208eac2d8d5f73d18d9eee99783500981013cb661"} Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.866420 4888 scope.go:117] "RemoveContainer" containerID="d8ec94040fb07cbf993713b53167ae2733ef9c9f0b35bd0f23df9a8f032b30e2" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.866158 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.890231 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.901575 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.908209 4888 scope.go:117] "RemoveContainer" containerID="515c450f600bede2918ee2a441c7b01a06083340cfbedabf222a57829ee5da1d" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.930316 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:08 crc kubenswrapper[4888]: E1124 01:40:08.930849 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="probe" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.930873 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="probe" Nov 24 01:40:08 crc kubenswrapper[4888]: E1124 01:40:08.930887 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="manila-share" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.930896 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="manila-share" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.931149 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="probe" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.931180 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" containerName="manila-share" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.932782 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.936873 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 01:40:08 crc kubenswrapper[4888]: I1124 01:40:08.943125 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.120892 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-scripts\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121012 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121045 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121432 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sd8m\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-kube-api-access-7sd8m\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121671 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121763 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-ceph\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.121884 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.122104 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.224935 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225034 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-ceph\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225087 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225208 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225294 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-scripts\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225418 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225481 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225408 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225641 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sd8m\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-kube-api-access-7sd8m\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.225675 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5f42ea1-62ff-4f74-bf41-4425b7332de8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.230332 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.235054 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.235490 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-scripts\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.235996 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f42ea1-62ff-4f74-bf41-4425b7332de8-config-data\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.236626 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-ceph\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.247184 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:40:09 crc kubenswrapper[4888]: E1124 01:40:09.247532 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.248951 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sd8m\" (UniqueName: \"kubernetes.io/projected/a5f42ea1-62ff-4f74-bf41-4425b7332de8-kube-api-access-7sd8m\") pod \"manila-share-share1-0\" (UID: \"a5f42ea1-62ff-4f74-bf41-4425b7332de8\") " pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.266012 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 01:40:09 crc kubenswrapper[4888]: I1124 01:40:09.910745 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 01:40:10 crc kubenswrapper[4888]: I1124 01:40:10.263222 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65746f6a-e726-41e1-adb1-33d86a5ad858" path="/var/lib/kubelet/pods/65746f6a-e726-41e1-adb1-33d86a5ad858/volumes" Nov 24 01:40:10 crc kubenswrapper[4888]: I1124 01:40:10.890137 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5f42ea1-62ff-4f74-bf41-4425b7332de8","Type":"ContainerStarted","Data":"84cd9641a5373b2a5d3e6d856dee58b7ef2117cc8891d53ccadbd4bee9c3acf9"} Nov 24 01:40:10 crc kubenswrapper[4888]: I1124 01:40:10.890418 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5f42ea1-62ff-4f74-bf41-4425b7332de8","Type":"ContainerStarted","Data":"a60b5dc4e264d521e8fcdc1b834d10685cf62d8953a15f523df57152ad12f285"} Nov 24 01:40:11 crc kubenswrapper[4888]: I1124 01:40:11.904920 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5f42ea1-62ff-4f74-bf41-4425b7332de8","Type":"ContainerStarted","Data":"a22b00ff70f1aa1969d8fcd2dd2db51881fd951719ce8f807e806d2aefdbdd83"} Nov 24 01:40:11 crc kubenswrapper[4888]: I1124 01:40:11.940706 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.9406839270000003 podStartE2EDuration="3.940683927s" podCreationTimestamp="2025-11-24 01:40:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:40:11.923656781 +0000 UTC m=+4514.506340875" watchObservedRunningTime="2025-11-24 01:40:11.940683927 +0000 UTC m=+4514.523367971" Nov 24 01:40:12 crc kubenswrapper[4888]: I1124 01:40:12.518796 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 01:40:13 crc kubenswrapper[4888]: I1124 01:40:13.645586 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776b89d8d8-zqnfp" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.68:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.68:8443: connect: connection refused" Nov 24 01:40:13 crc kubenswrapper[4888]: I1124 01:40:13.646132 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:40:18 crc kubenswrapper[4888]: I1124 01:40:18.986839 4888 generic.go:334] "Generic (PLEG): container finished" podID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerID="af15b819f8a3240dea553f709b65bef99f9954d0fae76056e04ecb24ffb06fe4" exitCode=137 Nov 24 01:40:18 crc kubenswrapper[4888]: I1124 01:40:18.987266 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerDied","Data":"af15b819f8a3240dea553f709b65bef99f9954d0fae76056e04ecb24ffb06fe4"} Nov 24 01:40:18 crc kubenswrapper[4888]: I1124 01:40:18.987294 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776b89d8d8-zqnfp" event={"ID":"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0","Type":"ContainerDied","Data":"06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3"} Nov 24 01:40:18 crc kubenswrapper[4888]: I1124 01:40:18.987304 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06e91041c6a641b64026faee3078bf3c28a0d2e279e6ed3d7e7873b6e01ab1b3" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.070324 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182145 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182230 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182277 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182347 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182389 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182414 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.182525 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqqc4\" (UniqueName: \"kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4\") pod \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\" (UID: \"fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0\") " Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.183250 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs" (OuterVolumeSpecName: "logs") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.191708 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.191800 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4" (OuterVolumeSpecName: "kube-api-access-pqqc4") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "kube-api-access-pqqc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.214746 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data" (OuterVolumeSpecName: "config-data") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.217051 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.233596 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts" (OuterVolumeSpecName: "scripts") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.244235 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" (UID: "fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.266891 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285553 4888 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285596 4888 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285615 4888 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-logs\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285633 4888 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285650 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285666 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.285683 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqqc4\" (UniqueName: \"kubernetes.io/projected/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0-kube-api-access-pqqc4\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:19 crc kubenswrapper[4888]: I1124 01:40:19.997650 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776b89d8d8-zqnfp" Nov 24 01:40:20 crc kubenswrapper[4888]: I1124 01:40:20.041089 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:40:20 crc kubenswrapper[4888]: I1124 01:40:20.052023 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-776b89d8d8-zqnfp"] Nov 24 01:40:20 crc kubenswrapper[4888]: I1124 01:40:20.258376 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" path="/var/lib/kubelet/pods/fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0/volumes" Nov 24 01:40:21 crc kubenswrapper[4888]: I1124 01:40:21.245296 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:40:21 crc kubenswrapper[4888]: E1124 01:40:21.246166 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:40:23 crc kubenswrapper[4888]: I1124 01:40:23.986123 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 01:40:28 crc kubenswrapper[4888]: I1124 01:40:28.171420 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.304597 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:29 crc kubenswrapper[4888]: E1124 01:40:29.305489 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.305507 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" Nov 24 01:40:29 crc kubenswrapper[4888]: E1124 01:40:29.305532 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon-log" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.305540 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon-log" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.305780 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon-log" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.305799 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5e39f0-8f83-4d1a-93d9-ed8d4e713ce0" containerName="horizon" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.307765 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.322476 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.423268 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.423453 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2dc\" (UniqueName: \"kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.423496 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.527714 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.527874 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2dc\" (UniqueName: \"kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.527916 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.528292 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.528444 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.553538 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2dc\" (UniqueName: \"kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc\") pod \"certified-operators-7ht67\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:29 crc kubenswrapper[4888]: I1124 01:40:29.637225 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:30 crc kubenswrapper[4888]: I1124 01:40:30.277436 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:30 crc kubenswrapper[4888]: I1124 01:40:30.801099 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 01:40:31 crc kubenswrapper[4888]: I1124 01:40:31.128275 4888 generic.go:334] "Generic (PLEG): container finished" podID="66606b3a-3c10-4bd9-9301-45d02859c616" containerID="d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d" exitCode=0 Nov 24 01:40:31 crc kubenswrapper[4888]: I1124 01:40:31.128324 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerDied","Data":"d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d"} Nov 24 01:40:31 crc kubenswrapper[4888]: I1124 01:40:31.128354 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerStarted","Data":"06756d579be0532cfbc25e0acd063d2b713d567cf8fa844b4ed50b0b00cd5aae"} Nov 24 01:40:33 crc kubenswrapper[4888]: I1124 01:40:33.153619 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerStarted","Data":"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be"} Nov 24 01:40:34 crc kubenswrapper[4888]: I1124 01:40:34.167861 4888 generic.go:334] "Generic (PLEG): container finished" podID="66606b3a-3c10-4bd9-9301-45d02859c616" containerID="278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be" exitCode=0 Nov 24 01:40:34 crc kubenswrapper[4888]: I1124 01:40:34.167988 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerDied","Data":"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be"} Nov 24 01:40:34 crc kubenswrapper[4888]: I1124 01:40:34.245446 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:40:34 crc kubenswrapper[4888]: E1124 01:40:34.245944 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:40:35 crc kubenswrapper[4888]: I1124 01:40:35.182802 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerStarted","Data":"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d"} Nov 24 01:40:35 crc kubenswrapper[4888]: I1124 01:40:35.221685 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7ht67" podStartSLOduration=2.635545457 podStartE2EDuration="6.221663238s" podCreationTimestamp="2025-11-24 01:40:29 +0000 UTC" firstStartedPulling="2025-11-24 01:40:31.131724493 +0000 UTC m=+4533.714408537" lastFinishedPulling="2025-11-24 01:40:34.717842244 +0000 UTC m=+4537.300526318" observedRunningTime="2025-11-24 01:40:35.21098225 +0000 UTC m=+4537.793666374" watchObservedRunningTime="2025-11-24 01:40:35.221663238 +0000 UTC m=+4537.804347292" Nov 24 01:40:39 crc kubenswrapper[4888]: I1124 01:40:39.638228 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:39 crc kubenswrapper[4888]: I1124 01:40:39.638788 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:39 crc kubenswrapper[4888]: I1124 01:40:39.713552 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:40 crc kubenswrapper[4888]: I1124 01:40:40.326603 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:41 crc kubenswrapper[4888]: I1124 01:40:41.071362 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.269118 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7ht67" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="registry-server" containerID="cri-o://b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d" gracePeriod=2 Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.845204 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.977187 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs2dc\" (UniqueName: \"kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc\") pod \"66606b3a-3c10-4bd9-9301-45d02859c616\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.977431 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content\") pod \"66606b3a-3c10-4bd9-9301-45d02859c616\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.987523 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc" (OuterVolumeSpecName: "kube-api-access-bs2dc") pod "66606b3a-3c10-4bd9-9301-45d02859c616" (UID: "66606b3a-3c10-4bd9-9301-45d02859c616"). InnerVolumeSpecName "kube-api-access-bs2dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.998654 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities\") pod \"66606b3a-3c10-4bd9-9301-45d02859c616\" (UID: \"66606b3a-3c10-4bd9-9301-45d02859c616\") " Nov 24 01:40:42 crc kubenswrapper[4888]: I1124 01:40:42.999400 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities" (OuterVolumeSpecName: "utilities") pod "66606b3a-3c10-4bd9-9301-45d02859c616" (UID: "66606b3a-3c10-4bd9-9301-45d02859c616"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.003318 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.003359 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs2dc\" (UniqueName: \"kubernetes.io/projected/66606b3a-3c10-4bd9-9301-45d02859c616-kube-api-access-bs2dc\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.045213 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66606b3a-3c10-4bd9-9301-45d02859c616" (UID: "66606b3a-3c10-4bd9-9301-45d02859c616"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.108314 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66606b3a-3c10-4bd9-9301-45d02859c616-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.288310 4888 generic.go:334] "Generic (PLEG): container finished" podID="66606b3a-3c10-4bd9-9301-45d02859c616" containerID="b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d" exitCode=0 Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.288356 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerDied","Data":"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d"} Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.288385 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ht67" event={"ID":"66606b3a-3c10-4bd9-9301-45d02859c616","Type":"ContainerDied","Data":"06756d579be0532cfbc25e0acd063d2b713d567cf8fa844b4ed50b0b00cd5aae"} Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.288403 4888 scope.go:117] "RemoveContainer" containerID="b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.288466 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ht67" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.335688 4888 scope.go:117] "RemoveContainer" containerID="278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be" Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.357418 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:43 crc kubenswrapper[4888]: I1124 01:40:43.377952 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7ht67"] Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.266542 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" path="/var/lib/kubelet/pods/66606b3a-3c10-4bd9-9301-45d02859c616/volumes" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.369543 4888 scope.go:117] "RemoveContainer" containerID="d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.427136 4888 scope.go:117] "RemoveContainer" containerID="b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d" Nov 24 01:40:44 crc kubenswrapper[4888]: E1124 01:40:44.427729 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d\": container with ID starting with b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d not found: ID does not exist" containerID="b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.427771 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d"} err="failed to get container status \"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d\": rpc error: code = NotFound desc = could not find container \"b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d\": container with ID starting with b5549496288cf633071e90fc277cb454dd89210709257e86ce6f916e622afb4d not found: ID does not exist" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.427797 4888 scope.go:117] "RemoveContainer" containerID="278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be" Nov 24 01:40:44 crc kubenswrapper[4888]: E1124 01:40:44.428656 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be\": container with ID starting with 278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be not found: ID does not exist" containerID="278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.428709 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be"} err="failed to get container status \"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be\": rpc error: code = NotFound desc = could not find container \"278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be\": container with ID starting with 278ac9eb1e482450471f5dcffab905b9f6795f4b19d401ad2e2cc9876b83c0be not found: ID does not exist" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.428736 4888 scope.go:117] "RemoveContainer" containerID="d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d" Nov 24 01:40:44 crc kubenswrapper[4888]: E1124 01:40:44.429114 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d\": container with ID starting with d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d not found: ID does not exist" containerID="d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d" Nov 24 01:40:44 crc kubenswrapper[4888]: I1124 01:40:44.429143 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d"} err="failed to get container status \"d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d\": rpc error: code = NotFound desc = could not find container \"d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d\": container with ID starting with d1e860765abdd82f465c3b962657f41f44b1971c5c2dca21aece6001c4369e0d not found: ID does not exist" Nov 24 01:40:47 crc kubenswrapper[4888]: I1124 01:40:47.246462 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:40:47 crc kubenswrapper[4888]: E1124 01:40:47.247589 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:41:01 crc kubenswrapper[4888]: I1124 01:41:01.246951 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:41:02 crc kubenswrapper[4888]: I1124 01:41:02.531385 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8"} Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.090110 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:31 crc kubenswrapper[4888]: E1124 01:41:31.091141 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="extract-utilities" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.091156 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="extract-utilities" Nov 24 01:41:31 crc kubenswrapper[4888]: E1124 01:41:31.091168 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="registry-server" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.091174 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="registry-server" Nov 24 01:41:31 crc kubenswrapper[4888]: E1124 01:41:31.091194 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="extract-content" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.091200 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="extract-content" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.091403 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="66606b3a-3c10-4bd9-9301-45d02859c616" containerName="registry-server" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.093043 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.106457 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.180249 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjc9g\" (UniqueName: \"kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.180488 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.180558 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.282353 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjc9g\" (UniqueName: \"kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.282548 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.282586 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.283090 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.283122 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.558229 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjc9g\" (UniqueName: \"kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g\") pod \"community-operators-f47bs\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:31 crc kubenswrapper[4888]: I1124 01:41:31.738183 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:32 crc kubenswrapper[4888]: I1124 01:41:32.382226 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:32 crc kubenswrapper[4888]: W1124 01:41:32.395432 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b65a366_5989_4a7f_9ead_1472fca549b2.slice/crio-ffa8b5f995128676af81c9e0986ff3fdd2617ea5f2ed7cef590de031be0e3024 WatchSource:0}: Error finding container ffa8b5f995128676af81c9e0986ff3fdd2617ea5f2ed7cef590de031be0e3024: Status 404 returned error can't find the container with id ffa8b5f995128676af81c9e0986ff3fdd2617ea5f2ed7cef590de031be0e3024 Nov 24 01:41:32 crc kubenswrapper[4888]: I1124 01:41:32.926415 4888 generic.go:334] "Generic (PLEG): container finished" podID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerID="8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835" exitCode=0 Nov 24 01:41:32 crc kubenswrapper[4888]: I1124 01:41:32.926476 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerDied","Data":"8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835"} Nov 24 01:41:32 crc kubenswrapper[4888]: I1124 01:41:32.927218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerStarted","Data":"ffa8b5f995128676af81c9e0986ff3fdd2617ea5f2ed7cef590de031be0e3024"} Nov 24 01:41:32 crc kubenswrapper[4888]: I1124 01:41:32.928686 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:41:33 crc kubenswrapper[4888]: I1124 01:41:33.942292 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerStarted","Data":"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1"} Nov 24 01:41:35 crc kubenswrapper[4888]: I1124 01:41:35.967709 4888 generic.go:334] "Generic (PLEG): container finished" podID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerID="0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1" exitCode=0 Nov 24 01:41:35 crc kubenswrapper[4888]: I1124 01:41:35.967770 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerDied","Data":"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1"} Nov 24 01:41:37 crc kubenswrapper[4888]: I1124 01:41:37.045244 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerStarted","Data":"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3"} Nov 24 01:41:37 crc kubenswrapper[4888]: I1124 01:41:37.073539 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f47bs" podStartSLOduration=2.606044756 podStartE2EDuration="6.073512221s" podCreationTimestamp="2025-11-24 01:41:31 +0000 UTC" firstStartedPulling="2025-11-24 01:41:32.928412573 +0000 UTC m=+4595.511096627" lastFinishedPulling="2025-11-24 01:41:36.395880018 +0000 UTC m=+4598.978564092" observedRunningTime="2025-11-24 01:41:37.070489516 +0000 UTC m=+4599.653173570" watchObservedRunningTime="2025-11-24 01:41:37.073512221 +0000 UTC m=+4599.656196285" Nov 24 01:41:41 crc kubenswrapper[4888]: I1124 01:41:41.739105 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:41 crc kubenswrapper[4888]: I1124 01:41:41.739707 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:41 crc kubenswrapper[4888]: I1124 01:41:41.832858 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:42 crc kubenswrapper[4888]: I1124 01:41:42.178982 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:42 crc kubenswrapper[4888]: I1124 01:41:42.243375 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.133001 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f47bs" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="registry-server" containerID="cri-o://c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3" gracePeriod=2 Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.669519 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.833204 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjc9g\" (UniqueName: \"kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g\") pod \"8b65a366-5989-4a7f-9ead-1472fca549b2\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.833877 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content\") pod \"8b65a366-5989-4a7f-9ead-1472fca549b2\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.834024 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities\") pod \"8b65a366-5989-4a7f-9ead-1472fca549b2\" (UID: \"8b65a366-5989-4a7f-9ead-1472fca549b2\") " Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.835432 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities" (OuterVolumeSpecName: "utilities") pod "8b65a366-5989-4a7f-9ead-1472fca549b2" (UID: "8b65a366-5989-4a7f-9ead-1472fca549b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.847621 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g" (OuterVolumeSpecName: "kube-api-access-rjc9g") pod "8b65a366-5989-4a7f-9ead-1472fca549b2" (UID: "8b65a366-5989-4a7f-9ead-1472fca549b2"). InnerVolumeSpecName "kube-api-access-rjc9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.887912 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b65a366-5989-4a7f-9ead-1472fca549b2" (UID: "8b65a366-5989-4a7f-9ead-1472fca549b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.936632 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.936667 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b65a366-5989-4a7f-9ead-1472fca549b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:41:44 crc kubenswrapper[4888]: I1124 01:41:44.936676 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjc9g\" (UniqueName: \"kubernetes.io/projected/8b65a366-5989-4a7f-9ead-1472fca549b2-kube-api-access-rjc9g\") on node \"crc\" DevicePath \"\"" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.151985 4888 generic.go:334] "Generic (PLEG): container finished" podID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerID="c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3" exitCode=0 Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.152046 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerDied","Data":"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3"} Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.152084 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f47bs" event={"ID":"8b65a366-5989-4a7f-9ead-1472fca549b2","Type":"ContainerDied","Data":"ffa8b5f995128676af81c9e0986ff3fdd2617ea5f2ed7cef590de031be0e3024"} Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.152118 4888 scope.go:117] "RemoveContainer" containerID="c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.152107 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f47bs" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.193420 4888 scope.go:117] "RemoveContainer" containerID="0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.204190 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.214308 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f47bs"] Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.219594 4888 scope.go:117] "RemoveContainer" containerID="8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.274605 4888 scope.go:117] "RemoveContainer" containerID="c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3" Nov 24 01:41:45 crc kubenswrapper[4888]: E1124 01:41:45.275152 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3\": container with ID starting with c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3 not found: ID does not exist" containerID="c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.275210 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3"} err="failed to get container status \"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3\": rpc error: code = NotFound desc = could not find container \"c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3\": container with ID starting with c46acb8c2d803cd56ac85f7499da35d52be02e3da73208670e4434e49c16f0f3 not found: ID does not exist" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.275240 4888 scope.go:117] "RemoveContainer" containerID="0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1" Nov 24 01:41:45 crc kubenswrapper[4888]: E1124 01:41:45.275545 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1\": container with ID starting with 0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1 not found: ID does not exist" containerID="0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.275570 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1"} err="failed to get container status \"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1\": rpc error: code = NotFound desc = could not find container \"0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1\": container with ID starting with 0df25a5c920fee88d6de6ebd444322e9a0a6b9976feed53bb5da06c9295c15f1 not found: ID does not exist" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.275585 4888 scope.go:117] "RemoveContainer" containerID="8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835" Nov 24 01:41:45 crc kubenswrapper[4888]: E1124 01:41:45.275940 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835\": container with ID starting with 8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835 not found: ID does not exist" containerID="8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835" Nov 24 01:41:45 crc kubenswrapper[4888]: I1124 01:41:45.275981 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835"} err="failed to get container status \"8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835\": rpc error: code = NotFound desc = could not find container \"8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835\": container with ID starting with 8ccdb5ddd47abe35dbcbb271457d1ddd3d9a911ceb50d97b5d7f557f9a2ae835 not found: ID does not exist" Nov 24 01:41:46 crc kubenswrapper[4888]: I1124 01:41:46.263314 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" path="/var/lib/kubelet/pods/8b65a366-5989-4a7f-9ead-1472fca549b2/volumes" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.589373 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:28 crc kubenswrapper[4888]: E1124 01:42:28.590480 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="registry-server" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.590494 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="registry-server" Nov 24 01:42:28 crc kubenswrapper[4888]: E1124 01:42:28.590527 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="extract-content" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.590534 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="extract-content" Nov 24 01:42:28 crc kubenswrapper[4888]: E1124 01:42:28.590545 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="extract-utilities" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.590552 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="extract-utilities" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.590795 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b65a366-5989-4a7f-9ead-1472fca549b2" containerName="registry-server" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.592438 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.625837 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.736337 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.736439 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.736661 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78xld\" (UniqueName: \"kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.838105 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.838245 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78xld\" (UniqueName: \"kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.838350 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.838552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.838781 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.855438 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78xld\" (UniqueName: \"kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld\") pod \"redhat-marketplace-nbjhv\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:28 crc kubenswrapper[4888]: I1124 01:42:28.938861 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:29 crc kubenswrapper[4888]: I1124 01:42:29.369489 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:29 crc kubenswrapper[4888]: I1124 01:42:29.746995 4888 generic.go:334] "Generic (PLEG): container finished" podID="49749af0-9390-48fb-abf6-04c868829f3b" containerID="cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29" exitCode=0 Nov 24 01:42:29 crc kubenswrapper[4888]: I1124 01:42:29.747054 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerDied","Data":"cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29"} Nov 24 01:42:29 crc kubenswrapper[4888]: I1124 01:42:29.748166 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerStarted","Data":"6000bfc7ea56ee78fc248bfa3243a637c5e35ec87c9c2b3b1c32f8ba3abfb36d"} Nov 24 01:42:31 crc kubenswrapper[4888]: I1124 01:42:31.773436 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerStarted","Data":"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6"} Nov 24 01:42:32 crc kubenswrapper[4888]: I1124 01:42:32.786383 4888 generic.go:334] "Generic (PLEG): container finished" podID="49749af0-9390-48fb-abf6-04c868829f3b" containerID="e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6" exitCode=0 Nov 24 01:42:32 crc kubenswrapper[4888]: I1124 01:42:32.786461 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerDied","Data":"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6"} Nov 24 01:42:33 crc kubenswrapper[4888]: I1124 01:42:33.801300 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerStarted","Data":"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349"} Nov 24 01:42:33 crc kubenswrapper[4888]: I1124 01:42:33.838581 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nbjhv" podStartSLOduration=2.39363199 podStartE2EDuration="5.83856403s" podCreationTimestamp="2025-11-24 01:42:28 +0000 UTC" firstStartedPulling="2025-11-24 01:42:29.748829624 +0000 UTC m=+4652.331513668" lastFinishedPulling="2025-11-24 01:42:33.193761624 +0000 UTC m=+4655.776445708" observedRunningTime="2025-11-24 01:42:33.83139194 +0000 UTC m=+4656.414075994" watchObservedRunningTime="2025-11-24 01:42:33.83856403 +0000 UTC m=+4656.421248074" Nov 24 01:42:38 crc kubenswrapper[4888]: I1124 01:42:38.939430 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:38 crc kubenswrapper[4888]: I1124 01:42:38.939790 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:39 crc kubenswrapper[4888]: I1124 01:42:39.912577 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:39 crc kubenswrapper[4888]: I1124 01:42:39.982593 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:40 crc kubenswrapper[4888]: I1124 01:42:40.165025 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:41 crc kubenswrapper[4888]: I1124 01:42:41.894628 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nbjhv" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="registry-server" containerID="cri-o://a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349" gracePeriod=2 Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.464144 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.579178 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities\") pod \"49749af0-9390-48fb-abf6-04c868829f3b\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.579372 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content\") pod \"49749af0-9390-48fb-abf6-04c868829f3b\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.579564 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78xld\" (UniqueName: \"kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld\") pod \"49749af0-9390-48fb-abf6-04c868829f3b\" (UID: \"49749af0-9390-48fb-abf6-04c868829f3b\") " Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.580253 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities" (OuterVolumeSpecName: "utilities") pod "49749af0-9390-48fb-abf6-04c868829f3b" (UID: "49749af0-9390-48fb-abf6-04c868829f3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.585526 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld" (OuterVolumeSpecName: "kube-api-access-78xld") pod "49749af0-9390-48fb-abf6-04c868829f3b" (UID: "49749af0-9390-48fb-abf6-04c868829f3b"). InnerVolumeSpecName "kube-api-access-78xld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.597101 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49749af0-9390-48fb-abf6-04c868829f3b" (UID: "49749af0-9390-48fb-abf6-04c868829f3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.682274 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.682313 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78xld\" (UniqueName: \"kubernetes.io/projected/49749af0-9390-48fb-abf6-04c868829f3b-kube-api-access-78xld\") on node \"crc\" DevicePath \"\"" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.682323 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49749af0-9390-48fb-abf6-04c868829f3b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.907702 4888 generic.go:334] "Generic (PLEG): container finished" podID="49749af0-9390-48fb-abf6-04c868829f3b" containerID="a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349" exitCode=0 Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.907743 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbjhv" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.907748 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerDied","Data":"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349"} Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.907777 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbjhv" event={"ID":"49749af0-9390-48fb-abf6-04c868829f3b","Type":"ContainerDied","Data":"6000bfc7ea56ee78fc248bfa3243a637c5e35ec87c9c2b3b1c32f8ba3abfb36d"} Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.907797 4888 scope.go:117] "RemoveContainer" containerID="a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.941756 4888 scope.go:117] "RemoveContainer" containerID="e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6" Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.947449 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.957776 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbjhv"] Nov 24 01:42:42 crc kubenswrapper[4888]: I1124 01:42:42.968169 4888 scope.go:117] "RemoveContainer" containerID="cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.035597 4888 scope.go:117] "RemoveContainer" containerID="a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349" Nov 24 01:42:43 crc kubenswrapper[4888]: E1124 01:42:43.036182 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349\": container with ID starting with a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349 not found: ID does not exist" containerID="a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.036211 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349"} err="failed to get container status \"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349\": rpc error: code = NotFound desc = could not find container \"a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349\": container with ID starting with a5857281217937c973451cea23bcf2c1ee92c41c75cd5a33b820854a943cc349 not found: ID does not exist" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.036231 4888 scope.go:117] "RemoveContainer" containerID="e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6" Nov 24 01:42:43 crc kubenswrapper[4888]: E1124 01:42:43.036533 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6\": container with ID starting with e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6 not found: ID does not exist" containerID="e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.036641 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6"} err="failed to get container status \"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6\": rpc error: code = NotFound desc = could not find container \"e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6\": container with ID starting with e3e24fec513953bd849c7662c87338e2cf5622295ee362d77e1d6dc3a60a6cd6 not found: ID does not exist" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.036742 4888 scope.go:117] "RemoveContainer" containerID="cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29" Nov 24 01:42:43 crc kubenswrapper[4888]: E1124 01:42:43.037142 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29\": container with ID starting with cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29 not found: ID does not exist" containerID="cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29" Nov 24 01:42:43 crc kubenswrapper[4888]: I1124 01:42:43.037166 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29"} err="failed to get container status \"cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29\": rpc error: code = NotFound desc = could not find container \"cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29\": container with ID starting with cce288a61dbe1be9aadaa1ac5a0200ad1a2682f04a02d5749993acd66e6c7b29 not found: ID does not exist" Nov 24 01:42:44 crc kubenswrapper[4888]: I1124 01:42:44.267378 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49749af0-9390-48fb-abf6-04c868829f3b" path="/var/lib/kubelet/pods/49749af0-9390-48fb-abf6-04c868829f3b/volumes" Nov 24 01:43:23 crc kubenswrapper[4888]: I1124 01:43:23.591941 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:43:23 crc kubenswrapper[4888]: I1124 01:43:23.592609 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:43:53 crc kubenswrapper[4888]: I1124 01:43:53.592185 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:43:53 crc kubenswrapper[4888]: I1124 01:43:53.592986 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:44:23 crc kubenswrapper[4888]: I1124 01:44:23.591961 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:44:23 crc kubenswrapper[4888]: I1124 01:44:23.592636 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:44:23 crc kubenswrapper[4888]: I1124 01:44:23.592686 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:44:23 crc kubenswrapper[4888]: I1124 01:44:23.593516 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:44:23 crc kubenswrapper[4888]: I1124 01:44:23.593584 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8" gracePeriod=600 Nov 24 01:44:24 crc kubenswrapper[4888]: I1124 01:44:24.230979 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8"} Nov 24 01:44:24 crc kubenswrapper[4888]: I1124 01:44:24.231731 4888 scope.go:117] "RemoveContainer" containerID="3d2db129254c4e761874545a34975b375ef51febdf933e6f92b0a14486782829" Nov 24 01:44:24 crc kubenswrapper[4888]: I1124 01:44:24.232633 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8" exitCode=0 Nov 24 01:44:24 crc kubenswrapper[4888]: I1124 01:44:24.232705 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a"} Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.173745 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2"] Nov 24 01:45:00 crc kubenswrapper[4888]: E1124 01:45:00.174892 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="extract-utilities" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.174911 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="extract-utilities" Nov 24 01:45:00 crc kubenswrapper[4888]: E1124 01:45:00.174952 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="registry-server" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.174961 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="registry-server" Nov 24 01:45:00 crc kubenswrapper[4888]: E1124 01:45:00.174990 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="extract-content" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.174999 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="extract-content" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.175281 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="49749af0-9390-48fb-abf6-04c868829f3b" containerName="registry-server" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.176459 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.179009 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.183613 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.189949 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2"] Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.240162 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.240252 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zssxf\" (UniqueName: \"kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.240331 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.342838 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.343196 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zssxf\" (UniqueName: \"kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.343436 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.344089 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.351359 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.371129 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zssxf\" (UniqueName: \"kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf\") pod \"collect-profiles-29399145-5s2t2\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:00 crc kubenswrapper[4888]: I1124 01:45:00.498477 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:01 crc kubenswrapper[4888]: I1124 01:45:01.049724 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2"] Nov 24 01:45:01 crc kubenswrapper[4888]: I1124 01:45:01.694930 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" event={"ID":"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9","Type":"ContainerStarted","Data":"64504b20012dd03a180271541054768948565107bcd6881b8077bbe1a2b4159f"} Nov 24 01:45:01 crc kubenswrapper[4888]: I1124 01:45:01.695215 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" event={"ID":"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9","Type":"ContainerStarted","Data":"f6f5ebef5c0aa1058ca87435b69b7df3a359831704ced5dc3c7710bce5b223a1"} Nov 24 01:45:01 crc kubenswrapper[4888]: I1124 01:45:01.724116 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" podStartSLOduration=1.72407519 podStartE2EDuration="1.72407519s" podCreationTimestamp="2025-11-24 01:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 01:45:01.719103431 +0000 UTC m=+4804.301787485" watchObservedRunningTime="2025-11-24 01:45:01.72407519 +0000 UTC m=+4804.306759234" Nov 24 01:45:02 crc kubenswrapper[4888]: I1124 01:45:02.709553 4888 generic.go:334] "Generic (PLEG): container finished" podID="1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" containerID="64504b20012dd03a180271541054768948565107bcd6881b8077bbe1a2b4159f" exitCode=0 Nov 24 01:45:02 crc kubenswrapper[4888]: I1124 01:45:02.709801 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" event={"ID":"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9","Type":"ContainerDied","Data":"64504b20012dd03a180271541054768948565107bcd6881b8077bbe1a2b4159f"} Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.183026 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.246184 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume\") pod \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.246266 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zssxf\" (UniqueName: \"kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf\") pod \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.246352 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume\") pod \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\" (UID: \"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9\") " Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.247177 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume" (OuterVolumeSpecName: "config-volume") pod "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" (UID: "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.256870 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" (UID: "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.257440 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf" (OuterVolumeSpecName: "kube-api-access-zssxf") pod "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" (UID: "1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9"). InnerVolumeSpecName "kube-api-access-zssxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.350786 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zssxf\" (UniqueName: \"kubernetes.io/projected/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-kube-api-access-zssxf\") on node \"crc\" DevicePath \"\"" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.351329 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.351352 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.746328 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.746454 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2" event={"ID":"1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9","Type":"ContainerDied","Data":"f6f5ebef5c0aa1058ca87435b69b7df3a359831704ced5dc3c7710bce5b223a1"} Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.746526 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6f5ebef5c0aa1058ca87435b69b7df3a359831704ced5dc3c7710bce5b223a1" Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.816980 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n"] Nov 24 01:45:04 crc kubenswrapper[4888]: I1124 01:45:04.829444 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399100-7df6n"] Nov 24 01:45:06 crc kubenswrapper[4888]: I1124 01:45:06.263543 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d772d2-5f1c-4410-af5c-758f282527ea" path="/var/lib/kubelet/pods/38d772d2-5f1c-4410-af5c-758f282527ea/volumes" Nov 24 01:45:18 crc kubenswrapper[4888]: I1124 01:45:18.626021 4888 scope.go:117] "RemoveContainer" containerID="af15b819f8a3240dea553f709b65bef99f9954d0fae76056e04ecb24ffb06fe4" Nov 24 01:45:18 crc kubenswrapper[4888]: I1124 01:45:18.667410 4888 scope.go:117] "RemoveContainer" containerID="3750dca87ec36b88d65df28c6d4b598a5d00baab06246896039f6a4c1cfb599a" Nov 24 01:45:49 crc kubenswrapper[4888]: I1124 01:45:49.559017 4888 trace.go:236] Trace[1723229782]: "Calculate volume metrics of wal for pod openshift-logging/logging-loki-ingester-0" (24-Nov-2025 01:45:48.536) (total time: 1022ms): Nov 24 01:45:49 crc kubenswrapper[4888]: Trace[1723229782]: [1.022665078s] [1.022665078s] END Nov 24 01:46:19 crc kubenswrapper[4888]: I1124 01:46:19.542943 4888 scope.go:117] "RemoveContainer" containerID="f50539478ff08dc8df8a461a56c2e90ae0da10d157feb90d0e04f5dbcd342759" Nov 24 01:46:19 crc kubenswrapper[4888]: I1124 01:46:19.587112 4888 scope.go:117] "RemoveContainer" containerID="c241f05f63597ea7ddc5c5f59643721b337b65e85987cbf0c54851c79c296917" Nov 24 01:46:19 crc kubenswrapper[4888]: I1124 01:46:19.856194 4888 scope.go:117] "RemoveContainer" containerID="2aa1f9522922d5ae085428107ad51ef4234301638302bd5b3e2319f2a0813556" Nov 24 01:46:23 crc kubenswrapper[4888]: I1124 01:46:23.591982 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:46:23 crc kubenswrapper[4888]: I1124 01:46:23.592653 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:46:53 crc kubenswrapper[4888]: I1124 01:46:53.592309 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:46:53 crc kubenswrapper[4888]: I1124 01:46:53.593166 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.872824 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:46:55 crc kubenswrapper[4888]: E1124 01:46:55.874748 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" containerName="collect-profiles" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.874771 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" containerName="collect-profiles" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.875046 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" containerName="collect-profiles" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.876697 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.883459 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.886799 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56spp\" (UniqueName: \"kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.886906 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.887084 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.988729 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.988850 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56spp\" (UniqueName: \"kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.988891 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.989251 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:55 crc kubenswrapper[4888]: I1124 01:46:55.989293 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:56 crc kubenswrapper[4888]: I1124 01:46:56.009663 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56spp\" (UniqueName: \"kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp\") pod \"redhat-operators-fp7kc\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:56 crc kubenswrapper[4888]: I1124 01:46:56.211126 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:46:56 crc kubenswrapper[4888]: I1124 01:46:56.731713 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:46:57 crc kubenswrapper[4888]: I1124 01:46:57.123784 4888 generic.go:334] "Generic (PLEG): container finished" podID="6274eb33-8790-4c32-8284-f45aa090adf5" containerID="b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159" exitCode=0 Nov 24 01:46:57 crc kubenswrapper[4888]: I1124 01:46:57.123849 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerDied","Data":"b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159"} Nov 24 01:46:57 crc kubenswrapper[4888]: I1124 01:46:57.124220 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerStarted","Data":"d67f7b3ef40f03bf385a067b67f1d9d6b91825a24c242898e5f856f52334b179"} Nov 24 01:46:57 crc kubenswrapper[4888]: I1124 01:46:57.125855 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:46:58 crc kubenswrapper[4888]: I1124 01:46:58.145725 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerStarted","Data":"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb"} Nov 24 01:47:02 crc kubenswrapper[4888]: I1124 01:47:02.198040 4888 generic.go:334] "Generic (PLEG): container finished" podID="6274eb33-8790-4c32-8284-f45aa090adf5" containerID="1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb" exitCode=0 Nov 24 01:47:02 crc kubenswrapper[4888]: I1124 01:47:02.198117 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerDied","Data":"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb"} Nov 24 01:47:04 crc kubenswrapper[4888]: I1124 01:47:04.227016 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerStarted","Data":"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84"} Nov 24 01:47:04 crc kubenswrapper[4888]: I1124 01:47:04.255074 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fp7kc" podStartSLOduration=3.742120827 podStartE2EDuration="9.255052623s" podCreationTimestamp="2025-11-24 01:46:55 +0000 UTC" firstStartedPulling="2025-11-24 01:46:57.125624259 +0000 UTC m=+4919.708308303" lastFinishedPulling="2025-11-24 01:47:02.638556045 +0000 UTC m=+4925.221240099" observedRunningTime="2025-11-24 01:47:04.254135167 +0000 UTC m=+4926.836819251" watchObservedRunningTime="2025-11-24 01:47:04.255052623 +0000 UTC m=+4926.837736677" Nov 24 01:47:06 crc kubenswrapper[4888]: I1124 01:47:06.211359 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:06 crc kubenswrapper[4888]: I1124 01:47:06.212014 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:07 crc kubenswrapper[4888]: I1124 01:47:07.278972 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fp7kc" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="registry-server" probeResult="failure" output=< Nov 24 01:47:07 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 01:47:07 crc kubenswrapper[4888]: > Nov 24 01:47:16 crc kubenswrapper[4888]: I1124 01:47:16.281372 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:16 crc kubenswrapper[4888]: I1124 01:47:16.348568 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:16 crc kubenswrapper[4888]: I1124 01:47:16.532646 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.362127 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fp7kc" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="registry-server" containerID="cri-o://2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84" gracePeriod=2 Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.862375 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.947567 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities\") pod \"6274eb33-8790-4c32-8284-f45aa090adf5\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.947610 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56spp\" (UniqueName: \"kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp\") pod \"6274eb33-8790-4c32-8284-f45aa090adf5\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.947667 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content\") pod \"6274eb33-8790-4c32-8284-f45aa090adf5\" (UID: \"6274eb33-8790-4c32-8284-f45aa090adf5\") " Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.948524 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities" (OuterVolumeSpecName: "utilities") pod "6274eb33-8790-4c32-8284-f45aa090adf5" (UID: "6274eb33-8790-4c32-8284-f45aa090adf5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:47:17 crc kubenswrapper[4888]: I1124 01:47:17.954846 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp" (OuterVolumeSpecName: "kube-api-access-56spp") pod "6274eb33-8790-4c32-8284-f45aa090adf5" (UID: "6274eb33-8790-4c32-8284-f45aa090adf5"). InnerVolumeSpecName "kube-api-access-56spp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.050343 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.050387 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56spp\" (UniqueName: \"kubernetes.io/projected/6274eb33-8790-4c32-8284-f45aa090adf5-kube-api-access-56spp\") on node \"crc\" DevicePath \"\"" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.057163 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6274eb33-8790-4c32-8284-f45aa090adf5" (UID: "6274eb33-8790-4c32-8284-f45aa090adf5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.153456 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6274eb33-8790-4c32-8284-f45aa090adf5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.374425 4888 generic.go:334] "Generic (PLEG): container finished" podID="6274eb33-8790-4c32-8284-f45aa090adf5" containerID="2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84" exitCode=0 Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.374495 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp7kc" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.374537 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerDied","Data":"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84"} Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.375542 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp7kc" event={"ID":"6274eb33-8790-4c32-8284-f45aa090adf5","Type":"ContainerDied","Data":"d67f7b3ef40f03bf385a067b67f1d9d6b91825a24c242898e5f856f52334b179"} Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.375568 4888 scope.go:117] "RemoveContainer" containerID="2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.403148 4888 scope.go:117] "RemoveContainer" containerID="1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.409438 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.420487 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fp7kc"] Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.431037 4888 scope.go:117] "RemoveContainer" containerID="b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.473099 4888 scope.go:117] "RemoveContainer" containerID="2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84" Nov 24 01:47:18 crc kubenswrapper[4888]: E1124 01:47:18.475153 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84\": container with ID starting with 2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84 not found: ID does not exist" containerID="2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.475202 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84"} err="failed to get container status \"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84\": rpc error: code = NotFound desc = could not find container \"2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84\": container with ID starting with 2735bfe85db23443646d8982d5ea596ce382d66caf86f331184d82fd9d252d84 not found: ID does not exist" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.475230 4888 scope.go:117] "RemoveContainer" containerID="1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb" Nov 24 01:47:18 crc kubenswrapper[4888]: E1124 01:47:18.475683 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb\": container with ID starting with 1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb not found: ID does not exist" containerID="1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.475714 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb"} err="failed to get container status \"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb\": rpc error: code = NotFound desc = could not find container \"1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb\": container with ID starting with 1508ca6b3fa17d7427abb5f0ff41bb2a46ed5a604bd9f1cece9a8cb6a99d39cb not found: ID does not exist" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.475735 4888 scope.go:117] "RemoveContainer" containerID="b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159" Nov 24 01:47:18 crc kubenswrapper[4888]: E1124 01:47:18.476066 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159\": container with ID starting with b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159 not found: ID does not exist" containerID="b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159" Nov 24 01:47:18 crc kubenswrapper[4888]: I1124 01:47:18.476148 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159"} err="failed to get container status \"b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159\": rpc error: code = NotFound desc = could not find container \"b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159\": container with ID starting with b76b86b1ba2f3fce9573025f1291c6ca4a3d30e34930a8790c0ade962ddec159 not found: ID does not exist" Nov 24 01:47:20 crc kubenswrapper[4888]: I1124 01:47:20.260935 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" path="/var/lib/kubelet/pods/6274eb33-8790-4c32-8284-f45aa090adf5/volumes" Nov 24 01:47:23 crc kubenswrapper[4888]: I1124 01:47:23.591561 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:47:23 crc kubenswrapper[4888]: I1124 01:47:23.592078 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:47:23 crc kubenswrapper[4888]: I1124 01:47:23.592122 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:47:23 crc kubenswrapper[4888]: I1124 01:47:23.592846 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:47:23 crc kubenswrapper[4888]: I1124 01:47:23.592892 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" gracePeriod=600 Nov 24 01:47:23 crc kubenswrapper[4888]: E1124 01:47:23.733519 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:47:24 crc kubenswrapper[4888]: I1124 01:47:24.446244 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" exitCode=0 Nov 24 01:47:24 crc kubenswrapper[4888]: I1124 01:47:24.446347 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a"} Nov 24 01:47:24 crc kubenswrapper[4888]: I1124 01:47:24.446633 4888 scope.go:117] "RemoveContainer" containerID="e9cb3feb52665089475c0113677aeb9a350c5aa9bb29d5b98a1eba590da632b8" Nov 24 01:47:24 crc kubenswrapper[4888]: I1124 01:47:24.447418 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:47:24 crc kubenswrapper[4888]: E1124 01:47:24.447770 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:47:38 crc kubenswrapper[4888]: I1124 01:47:38.245911 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:47:38 crc kubenswrapper[4888]: E1124 01:47:38.246829 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:47:50 crc kubenswrapper[4888]: I1124 01:47:50.245209 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:47:50 crc kubenswrapper[4888]: E1124 01:47:50.246274 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:48:04 crc kubenswrapper[4888]: I1124 01:48:04.246174 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:48:04 crc kubenswrapper[4888]: E1124 01:48:04.247185 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:48:16 crc kubenswrapper[4888]: I1124 01:48:16.249533 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:48:16 crc kubenswrapper[4888]: E1124 01:48:16.250501 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:48:21 crc kubenswrapper[4888]: I1124 01:48:21.747343 4888 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="9a963d44-99cd-45d4-afd8-9c9c7bfb95e7" containerName="galera" probeResult="failure" output="command timed out" Nov 24 01:48:21 crc kubenswrapper[4888]: I1124 01:48:21.747344 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="9a963d44-99cd-45d4-afd8-9c9c7bfb95e7" containerName="galera" probeResult="failure" output="command timed out" Nov 24 01:48:30 crc kubenswrapper[4888]: I1124 01:48:30.249337 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:48:30 crc kubenswrapper[4888]: E1124 01:48:30.250547 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:48:44 crc kubenswrapper[4888]: I1124 01:48:44.246966 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:48:44 crc kubenswrapper[4888]: E1124 01:48:44.248069 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:48:56 crc kubenswrapper[4888]: I1124 01:48:56.246439 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:48:56 crc kubenswrapper[4888]: E1124 01:48:56.247571 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:49:05 crc kubenswrapper[4888]: I1124 01:49:05.092804 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-4j9db"] Nov 24 01:49:05 crc kubenswrapper[4888]: I1124 01:49:05.102959 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-129e-account-create-8fkkk"] Nov 24 01:49:05 crc kubenswrapper[4888]: I1124 01:49:05.112287 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-4j9db"] Nov 24 01:49:05 crc kubenswrapper[4888]: I1124 01:49:05.121002 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-129e-account-create-8fkkk"] Nov 24 01:49:06 crc kubenswrapper[4888]: I1124 01:49:06.261765 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a4de536-bef6-4f2b-ac46-64a43f970422" path="/var/lib/kubelet/pods/3a4de536-bef6-4f2b-ac46-64a43f970422/volumes" Nov 24 01:49:06 crc kubenswrapper[4888]: I1124 01:49:06.264984 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76da0164-f26f-445f-bc36-401a45149229" path="/var/lib/kubelet/pods/76da0164-f26f-445f-bc36-401a45149229/volumes" Nov 24 01:49:10 crc kubenswrapper[4888]: I1124 01:49:10.246922 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:49:10 crc kubenswrapper[4888]: E1124 01:49:10.249104 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:49:20 crc kubenswrapper[4888]: I1124 01:49:20.051652 4888 scope.go:117] "RemoveContainer" containerID="5eda77f7cea7914596e6924e60779310c774daad78688d70a610ac3ba1e32049" Nov 24 01:49:20 crc kubenswrapper[4888]: I1124 01:49:20.769455 4888 scope.go:117] "RemoveContainer" containerID="55e094c08ed9c3977990f5c6d7782536d16892a8ba97a266219b2d15611b3014" Nov 24 01:49:23 crc kubenswrapper[4888]: I1124 01:49:23.246423 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:49:23 crc kubenswrapper[4888]: E1124 01:49:23.247505 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:49:33 crc kubenswrapper[4888]: I1124 01:49:33.068201 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-xg7z9"] Nov 24 01:49:33 crc kubenswrapper[4888]: I1124 01:49:33.083404 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-xg7z9"] Nov 24 01:49:34 crc kubenswrapper[4888]: I1124 01:49:34.269508 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e08e0b-9532-4dee-92ff-10324d806d08" path="/var/lib/kubelet/pods/04e08e0b-9532-4dee-92ff-10324d806d08/volumes" Nov 24 01:49:38 crc kubenswrapper[4888]: I1124 01:49:38.252501 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:49:38 crc kubenswrapper[4888]: E1124 01:49:38.253452 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:49:50 crc kubenswrapper[4888]: I1124 01:49:50.246690 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:49:50 crc kubenswrapper[4888]: E1124 01:49:50.247478 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:03 crc kubenswrapper[4888]: I1124 01:50:03.245938 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:50:03 crc kubenswrapper[4888]: E1124 01:50:03.246682 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:14 crc kubenswrapper[4888]: I1124 01:50:14.245375 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:50:14 crc kubenswrapper[4888]: E1124 01:50:14.246693 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:20 crc kubenswrapper[4888]: I1124 01:50:20.916139 4888 scope.go:117] "RemoveContainer" containerID="d3c87ca81b5b759dd447e6d765e75144f5d70321af7010a070c54364d7844742" Nov 24 01:50:28 crc kubenswrapper[4888]: I1124 01:50:28.258010 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:50:28 crc kubenswrapper[4888]: E1124 01:50:28.259204 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.917273 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:35 crc kubenswrapper[4888]: E1124 01:50:35.918304 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="registry-server" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.918318 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="registry-server" Nov 24 01:50:35 crc kubenswrapper[4888]: E1124 01:50:35.918331 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="extract-utilities" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.918337 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="extract-utilities" Nov 24 01:50:35 crc kubenswrapper[4888]: E1124 01:50:35.918388 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="extract-content" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.918395 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="extract-content" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.918605 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="6274eb33-8790-4c32-8284-f45aa090adf5" containerName="registry-server" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.922660 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:35 crc kubenswrapper[4888]: I1124 01:50:35.939716 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.070454 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.070539 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdttw\" (UniqueName: \"kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.070736 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.173301 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.173538 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.173596 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdttw\" (UniqueName: \"kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.173887 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.173965 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.546102 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdttw\" (UniqueName: \"kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw\") pod \"certified-operators-fk75g\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:36 crc kubenswrapper[4888]: I1124 01:50:36.559196 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:37 crc kubenswrapper[4888]: I1124 01:50:37.148983 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:37 crc kubenswrapper[4888]: I1124 01:50:37.892464 4888 generic.go:334] "Generic (PLEG): container finished" podID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerID="75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd" exitCode=0 Nov 24 01:50:37 crc kubenswrapper[4888]: I1124 01:50:37.892912 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerDied","Data":"75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd"} Nov 24 01:50:37 crc kubenswrapper[4888]: I1124 01:50:37.893028 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerStarted","Data":"974c60cde3ee16ada9749e2aa8c3cc9bec37f8ebf1f59224eec450f4598f5a25"} Nov 24 01:50:38 crc kubenswrapper[4888]: I1124 01:50:38.908163 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerStarted","Data":"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611"} Nov 24 01:50:39 crc kubenswrapper[4888]: I1124 01:50:39.926044 4888 generic.go:334] "Generic (PLEG): container finished" podID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerID="49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611" exitCode=0 Nov 24 01:50:39 crc kubenswrapper[4888]: I1124 01:50:39.926180 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerDied","Data":"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611"} Nov 24 01:50:40 crc kubenswrapper[4888]: I1124 01:50:40.246248 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:50:40 crc kubenswrapper[4888]: E1124 01:50:40.246871 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:40 crc kubenswrapper[4888]: I1124 01:50:40.943487 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerStarted","Data":"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb"} Nov 24 01:50:40 crc kubenswrapper[4888]: I1124 01:50:40.975456 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fk75g" podStartSLOduration=3.506242755 podStartE2EDuration="5.975436674s" podCreationTimestamp="2025-11-24 01:50:35 +0000 UTC" firstStartedPulling="2025-11-24 01:50:37.895234595 +0000 UTC m=+5140.477918639" lastFinishedPulling="2025-11-24 01:50:40.364428504 +0000 UTC m=+5142.947112558" observedRunningTime="2025-11-24 01:50:40.974608861 +0000 UTC m=+5143.557292915" watchObservedRunningTime="2025-11-24 01:50:40.975436674 +0000 UTC m=+5143.558120728" Nov 24 01:50:46 crc kubenswrapper[4888]: I1124 01:50:46.559324 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:46 crc kubenswrapper[4888]: I1124 01:50:46.560193 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:47 crc kubenswrapper[4888]: I1124 01:50:47.308688 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:47 crc kubenswrapper[4888]: I1124 01:50:47.383437 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:47 crc kubenswrapper[4888]: I1124 01:50:47.555109 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:49 crc kubenswrapper[4888]: I1124 01:50:49.040442 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fk75g" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="registry-server" containerID="cri-o://c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb" gracePeriod=2 Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.025702 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.070547 4888 generic.go:334] "Generic (PLEG): container finished" podID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerID="c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb" exitCode=0 Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.070588 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerDied","Data":"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb"} Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.070611 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk75g" event={"ID":"2f273eae-140c-46e9-85a2-db3d7b41dae8","Type":"ContainerDied","Data":"974c60cde3ee16ada9749e2aa8c3cc9bec37f8ebf1f59224eec450f4598f5a25"} Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.070637 4888 scope.go:117] "RemoveContainer" containerID="c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.070834 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk75g" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.106285 4888 scope.go:117] "RemoveContainer" containerID="49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.117917 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content\") pod \"2f273eae-140c-46e9-85a2-db3d7b41dae8\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.118171 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdttw\" (UniqueName: \"kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw\") pod \"2f273eae-140c-46e9-85a2-db3d7b41dae8\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.118393 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities\") pod \"2f273eae-140c-46e9-85a2-db3d7b41dae8\" (UID: \"2f273eae-140c-46e9-85a2-db3d7b41dae8\") " Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.120320 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities" (OuterVolumeSpecName: "utilities") pod "2f273eae-140c-46e9-85a2-db3d7b41dae8" (UID: "2f273eae-140c-46e9-85a2-db3d7b41dae8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.128295 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw" (OuterVolumeSpecName: "kube-api-access-mdttw") pod "2f273eae-140c-46e9-85a2-db3d7b41dae8" (UID: "2f273eae-140c-46e9-85a2-db3d7b41dae8"). InnerVolumeSpecName "kube-api-access-mdttw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.135893 4888 scope.go:117] "RemoveContainer" containerID="75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.193054 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f273eae-140c-46e9-85a2-db3d7b41dae8" (UID: "2f273eae-140c-46e9-85a2-db3d7b41dae8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.221784 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.221849 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdttw\" (UniqueName: \"kubernetes.io/projected/2f273eae-140c-46e9-85a2-db3d7b41dae8-kube-api-access-mdttw\") on node \"crc\" DevicePath \"\"" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.221863 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f273eae-140c-46e9-85a2-db3d7b41dae8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.275057 4888 scope.go:117] "RemoveContainer" containerID="c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb" Nov 24 01:50:50 crc kubenswrapper[4888]: E1124 01:50:50.275578 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb\": container with ID starting with c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb not found: ID does not exist" containerID="c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.275619 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb"} err="failed to get container status \"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb\": rpc error: code = NotFound desc = could not find container \"c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb\": container with ID starting with c84a291dd38efd5fe1e83dfb2cb0a5ec328a5e0002612936cb458f391a5be7eb not found: ID does not exist" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.275639 4888 scope.go:117] "RemoveContainer" containerID="49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611" Nov 24 01:50:50 crc kubenswrapper[4888]: E1124 01:50:50.276436 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611\": container with ID starting with 49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611 not found: ID does not exist" containerID="49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.276458 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611"} err="failed to get container status \"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611\": rpc error: code = NotFound desc = could not find container \"49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611\": container with ID starting with 49994f44ec1fcd2dcb12dd40f1d372ef9503ed94f215ffe823b567eb53617611 not found: ID does not exist" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.276471 4888 scope.go:117] "RemoveContainer" containerID="75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd" Nov 24 01:50:50 crc kubenswrapper[4888]: E1124 01:50:50.277042 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd\": container with ID starting with 75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd not found: ID does not exist" containerID="75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.277088 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd"} err="failed to get container status \"75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd\": rpc error: code = NotFound desc = could not find container \"75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd\": container with ID starting with 75034ea539a6804282628b0f6a1658af6f70504670921f13d655f199b5c7a5cd not found: ID does not exist" Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.412347 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:50 crc kubenswrapper[4888]: I1124 01:50:50.422017 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fk75g"] Nov 24 01:50:51 crc kubenswrapper[4888]: I1124 01:50:51.246509 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:50:51 crc kubenswrapper[4888]: E1124 01:50:51.247502 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:50:52 crc kubenswrapper[4888]: I1124 01:50:52.257167 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" path="/var/lib/kubelet/pods/2f273eae-140c-46e9-85a2-db3d7b41dae8/volumes" Nov 24 01:51:06 crc kubenswrapper[4888]: I1124 01:51:06.245886 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:51:06 crc kubenswrapper[4888]: E1124 01:51:06.246714 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:51:19 crc kubenswrapper[4888]: I1124 01:51:19.245965 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:51:19 crc kubenswrapper[4888]: E1124 01:51:19.247158 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:51:33 crc kubenswrapper[4888]: I1124 01:51:33.247522 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:51:33 crc kubenswrapper[4888]: E1124 01:51:33.250968 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:51:46 crc kubenswrapper[4888]: I1124 01:51:46.246368 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:51:46 crc kubenswrapper[4888]: E1124 01:51:46.247282 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:51:59 crc kubenswrapper[4888]: I1124 01:51:59.245748 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:51:59 crc kubenswrapper[4888]: E1124 01:51:59.246760 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:52:10 crc kubenswrapper[4888]: I1124 01:52:10.246653 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:52:10 crc kubenswrapper[4888]: E1124 01:52:10.247887 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:52:22 crc kubenswrapper[4888]: I1124 01:52:22.246545 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:52:22 crc kubenswrapper[4888]: E1124 01:52:22.247808 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:52:35 crc kubenswrapper[4888]: I1124 01:52:35.246394 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:52:36 crc kubenswrapper[4888]: I1124 01:52:36.550185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23"} Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.697668 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:40 crc kubenswrapper[4888]: E1124 01:53:40.698924 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="extract-content" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.698943 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="extract-content" Nov 24 01:53:40 crc kubenswrapper[4888]: E1124 01:53:40.698993 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="extract-utilities" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.699003 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="extract-utilities" Nov 24 01:53:40 crc kubenswrapper[4888]: E1124 01:53:40.699020 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="registry-server" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.699029 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="registry-server" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.699333 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f273eae-140c-46e9-85a2-db3d7b41dae8" containerName="registry-server" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.701401 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.711279 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.826218 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.826371 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.826649 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr6ps\" (UniqueName: \"kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.928342 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr6ps\" (UniqueName: \"kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.928483 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.928575 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.929022 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.929065 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:40 crc kubenswrapper[4888]: I1124 01:53:40.950323 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr6ps\" (UniqueName: \"kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps\") pod \"redhat-marketplace-7jtvp\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:41 crc kubenswrapper[4888]: I1124 01:53:41.025177 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:41 crc kubenswrapper[4888]: I1124 01:53:41.520346 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:42 crc kubenswrapper[4888]: I1124 01:53:42.238521 4888 generic.go:334] "Generic (PLEG): container finished" podID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerID="d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399" exitCode=0 Nov 24 01:53:42 crc kubenswrapper[4888]: I1124 01:53:42.238913 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerDied","Data":"d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399"} Nov 24 01:53:42 crc kubenswrapper[4888]: I1124 01:53:42.241170 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerStarted","Data":"76d14d20c78c949421b08cd682bb2309e6559f7df179c4a227d2bc97df63dfd1"} Nov 24 01:53:42 crc kubenswrapper[4888]: I1124 01:53:42.241095 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 01:53:43 crc kubenswrapper[4888]: I1124 01:53:43.253255 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerStarted","Data":"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8"} Nov 24 01:53:44 crc kubenswrapper[4888]: I1124 01:53:44.270660 4888 generic.go:334] "Generic (PLEG): container finished" podID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerID="908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8" exitCode=0 Nov 24 01:53:44 crc kubenswrapper[4888]: I1124 01:53:44.271200 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerDied","Data":"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8"} Nov 24 01:53:45 crc kubenswrapper[4888]: I1124 01:53:45.289653 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerStarted","Data":"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81"} Nov 24 01:53:45 crc kubenswrapper[4888]: I1124 01:53:45.322289 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7jtvp" podStartSLOduration=2.902603541 podStartE2EDuration="5.322264009s" podCreationTimestamp="2025-11-24 01:53:40 +0000 UTC" firstStartedPulling="2025-11-24 01:53:42.24078802 +0000 UTC m=+5324.823472074" lastFinishedPulling="2025-11-24 01:53:44.660448478 +0000 UTC m=+5327.243132542" observedRunningTime="2025-11-24 01:53:45.321131207 +0000 UTC m=+5327.903815301" watchObservedRunningTime="2025-11-24 01:53:45.322264009 +0000 UTC m=+5327.904948093" Nov 24 01:53:51 crc kubenswrapper[4888]: I1124 01:53:51.026042 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:51 crc kubenswrapper[4888]: I1124 01:53:51.026848 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:51 crc kubenswrapper[4888]: I1124 01:53:51.105242 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:51 crc kubenswrapper[4888]: I1124 01:53:51.448737 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:51 crc kubenswrapper[4888]: I1124 01:53:51.509353 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:53 crc kubenswrapper[4888]: I1124 01:53:53.410653 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7jtvp" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="registry-server" containerID="cri-o://def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81" gracePeriod=2 Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.053078 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.237028 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities\") pod \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.237329 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content\") pod \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.237443 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr6ps\" (UniqueName: \"kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps\") pod \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\" (UID: \"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b\") " Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.237897 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities" (OuterVolumeSpecName: "utilities") pod "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" (UID: "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.238243 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.256416 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps" (OuterVolumeSpecName: "kube-api-access-kr6ps") pod "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" (UID: "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b"). InnerVolumeSpecName "kube-api-access-kr6ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.272190 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" (UID: "22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.342019 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.342091 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr6ps\" (UniqueName: \"kubernetes.io/projected/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b-kube-api-access-kr6ps\") on node \"crc\" DevicePath \"\"" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.431197 4888 generic.go:334] "Generic (PLEG): container finished" podID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerID="def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81" exitCode=0 Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.431256 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerDied","Data":"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81"} Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.431294 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jtvp" event={"ID":"22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b","Type":"ContainerDied","Data":"76d14d20c78c949421b08cd682bb2309e6559f7df179c4a227d2bc97df63dfd1"} Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.431326 4888 scope.go:117] "RemoveContainer" containerID="def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.431385 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jtvp" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.482580 4888 scope.go:117] "RemoveContainer" containerID="908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8" Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.489051 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.509471 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jtvp"] Nov 24 01:53:54 crc kubenswrapper[4888]: I1124 01:53:54.517374 4888 scope.go:117] "RemoveContainer" containerID="d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.374698 4888 scope.go:117] "RemoveContainer" containerID="def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81" Nov 24 01:53:55 crc kubenswrapper[4888]: E1124 01:53:55.376130 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81\": container with ID starting with def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81 not found: ID does not exist" containerID="def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.376184 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81"} err="failed to get container status \"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81\": rpc error: code = NotFound desc = could not find container \"def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81\": container with ID starting with def6fcc4e5db9279800b87b4d22824198ec2209b2dcf3f60208ddac89debdb81 not found: ID does not exist" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.376223 4888 scope.go:117] "RemoveContainer" containerID="908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8" Nov 24 01:53:55 crc kubenswrapper[4888]: E1124 01:53:55.376905 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8\": container with ID starting with 908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8 not found: ID does not exist" containerID="908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.377020 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8"} err="failed to get container status \"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8\": rpc error: code = NotFound desc = could not find container \"908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8\": container with ID starting with 908bdbbbc1ad0ca9db0dd5f64b25c90e1c37503528b831761aa0315067e6f1a8 not found: ID does not exist" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.377105 4888 scope.go:117] "RemoveContainer" containerID="d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399" Nov 24 01:53:55 crc kubenswrapper[4888]: E1124 01:53:55.377662 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399\": container with ID starting with d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399 not found: ID does not exist" containerID="d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399" Nov 24 01:53:55 crc kubenswrapper[4888]: I1124 01:53:55.377696 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399"} err="failed to get container status \"d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399\": rpc error: code = NotFound desc = could not find container \"d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399\": container with ID starting with d9a6c5d354a240b102773b743c65b6cdaa5304e91c8d0bb7fe78c48ab9d3a399 not found: ID does not exist" Nov 24 01:53:56 crc kubenswrapper[4888]: I1124 01:53:56.292565 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" path="/var/lib/kubelet/pods/22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b/volumes" Nov 24 01:54:53 crc kubenswrapper[4888]: I1124 01:54:53.592018 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:54:53 crc kubenswrapper[4888]: I1124 01:54:53.592728 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.881252 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:11 crc kubenswrapper[4888]: E1124 01:55:11.882283 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="extract-utilities" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.882301 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="extract-utilities" Nov 24 01:55:11 crc kubenswrapper[4888]: E1124 01:55:11.882314 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="registry-server" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.882322 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="registry-server" Nov 24 01:55:11 crc kubenswrapper[4888]: E1124 01:55:11.882341 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="extract-content" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.882349 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="extract-content" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.882690 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a49ce7-1cb7-4d0a-b390-6b4acfff7b5b" containerName="registry-server" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.884691 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.910063 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.933907 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmm2h\" (UniqueName: \"kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.934000 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:11 crc kubenswrapper[4888]: I1124 01:55:11.934041 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.036441 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmm2h\" (UniqueName: \"kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.036758 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.036917 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.037553 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.037665 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.065764 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmm2h\" (UniqueName: \"kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h\") pod \"community-operators-f2dxb\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.224105 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:12 crc kubenswrapper[4888]: I1124 01:55:12.823714 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:13 crc kubenswrapper[4888]: I1124 01:55:13.480213 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerStarted","Data":"bd5fbc171c1fcb8bbd1e9210942f8cc85d2a98f12f958c54ac08fdb937c9c9e5"} Nov 24 01:55:14 crc kubenswrapper[4888]: I1124 01:55:14.495161 4888 generic.go:334] "Generic (PLEG): container finished" podID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerID="d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe" exitCode=0 Nov 24 01:55:14 crc kubenswrapper[4888]: I1124 01:55:14.495261 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerDied","Data":"d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe"} Nov 24 01:55:15 crc kubenswrapper[4888]: I1124 01:55:15.506950 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerStarted","Data":"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c"} Nov 24 01:55:17 crc kubenswrapper[4888]: I1124 01:55:17.534119 4888 generic.go:334] "Generic (PLEG): container finished" podID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerID="a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c" exitCode=0 Nov 24 01:55:17 crc kubenswrapper[4888]: I1124 01:55:17.534287 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerDied","Data":"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c"} Nov 24 01:55:18 crc kubenswrapper[4888]: I1124 01:55:18.551681 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerStarted","Data":"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b"} Nov 24 01:55:18 crc kubenswrapper[4888]: I1124 01:55:18.591544 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f2dxb" podStartSLOduration=4.153086315 podStartE2EDuration="7.591518111s" podCreationTimestamp="2025-11-24 01:55:11 +0000 UTC" firstStartedPulling="2025-11-24 01:55:14.497270932 +0000 UTC m=+5417.079954976" lastFinishedPulling="2025-11-24 01:55:17.935702718 +0000 UTC m=+5420.518386772" observedRunningTime="2025-11-24 01:55:18.576428018 +0000 UTC m=+5421.159112102" watchObservedRunningTime="2025-11-24 01:55:18.591518111 +0000 UTC m=+5421.174202185" Nov 24 01:55:22 crc kubenswrapper[4888]: I1124 01:55:22.225358 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:22 crc kubenswrapper[4888]: I1124 01:55:22.226660 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:22 crc kubenswrapper[4888]: I1124 01:55:22.299285 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:23 crc kubenswrapper[4888]: I1124 01:55:23.592243 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:55:23 crc kubenswrapper[4888]: I1124 01:55:23.593044 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:55:32 crc kubenswrapper[4888]: I1124 01:55:32.323498 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:32 crc kubenswrapper[4888]: I1124 01:55:32.404415 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:32 crc kubenswrapper[4888]: I1124 01:55:32.742889 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f2dxb" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="registry-server" containerID="cri-o://d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b" gracePeriod=2 Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.316976 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.503941 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities\") pod \"501b7734-df4d-4c4f-9bd0-56994ef6c447\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.504002 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content\") pod \"501b7734-df4d-4c4f-9bd0-56994ef6c447\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.504257 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmm2h\" (UniqueName: \"kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h\") pod \"501b7734-df4d-4c4f-9bd0-56994ef6c447\" (UID: \"501b7734-df4d-4c4f-9bd0-56994ef6c447\") " Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.505274 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities" (OuterVolumeSpecName: "utilities") pod "501b7734-df4d-4c4f-9bd0-56994ef6c447" (UID: "501b7734-df4d-4c4f-9bd0-56994ef6c447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.510393 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h" (OuterVolumeSpecName: "kube-api-access-lmm2h") pod "501b7734-df4d-4c4f-9bd0-56994ef6c447" (UID: "501b7734-df4d-4c4f-9bd0-56994ef6c447"). InnerVolumeSpecName "kube-api-access-lmm2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.556732 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "501b7734-df4d-4c4f-9bd0-56994ef6c447" (UID: "501b7734-df4d-4c4f-9bd0-56994ef6c447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.606602 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.606639 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501b7734-df4d-4c4f-9bd0-56994ef6c447-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.606651 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmm2h\" (UniqueName: \"kubernetes.io/projected/501b7734-df4d-4c4f-9bd0-56994ef6c447-kube-api-access-lmm2h\") on node \"crc\" DevicePath \"\"" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.761274 4888 generic.go:334] "Generic (PLEG): container finished" podID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerID="d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b" exitCode=0 Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.761350 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerDied","Data":"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b"} Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.761404 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2dxb" event={"ID":"501b7734-df4d-4c4f-9bd0-56994ef6c447","Type":"ContainerDied","Data":"bd5fbc171c1fcb8bbd1e9210942f8cc85d2a98f12f958c54ac08fdb937c9c9e5"} Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.761439 4888 scope.go:117] "RemoveContainer" containerID="d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.761707 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2dxb" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.802514 4888 scope.go:117] "RemoveContainer" containerID="a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.825388 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.836888 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f2dxb"] Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.842254 4888 scope.go:117] "RemoveContainer" containerID="d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.907332 4888 scope.go:117] "RemoveContainer" containerID="d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b" Nov 24 01:55:33 crc kubenswrapper[4888]: E1124 01:55:33.909173 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b\": container with ID starting with d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b not found: ID does not exist" containerID="d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.909205 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b"} err="failed to get container status \"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b\": rpc error: code = NotFound desc = could not find container \"d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b\": container with ID starting with d9b8744794a6c66077e614d159a43c843086fa3261627e76db0495747fb2867b not found: ID does not exist" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.909228 4888 scope.go:117] "RemoveContainer" containerID="a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c" Nov 24 01:55:33 crc kubenswrapper[4888]: E1124 01:55:33.909551 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c\": container with ID starting with a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c not found: ID does not exist" containerID="a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.909572 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c"} err="failed to get container status \"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c\": rpc error: code = NotFound desc = could not find container \"a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c\": container with ID starting with a9d60943efcd10ab03a309b75c60ea83bb387e73db77a52059ccee890316046c not found: ID does not exist" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.909583 4888 scope.go:117] "RemoveContainer" containerID="d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe" Nov 24 01:55:33 crc kubenswrapper[4888]: E1124 01:55:33.909867 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe\": container with ID starting with d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe not found: ID does not exist" containerID="d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe" Nov 24 01:55:33 crc kubenswrapper[4888]: I1124 01:55:33.909890 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe"} err="failed to get container status \"d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe\": rpc error: code = NotFound desc = could not find container \"d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe\": container with ID starting with d7a3b1042feb6e7c24b87e5bef5a60789fa595a9fb7ce10dde6a4474c55860fe not found: ID does not exist" Nov 24 01:55:34 crc kubenswrapper[4888]: I1124 01:55:34.265866 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" path="/var/lib/kubelet/pods/501b7734-df4d-4c4f-9bd0-56994ef6c447/volumes" Nov 24 01:55:53 crc kubenswrapper[4888]: I1124 01:55:53.591852 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:55:53 crc kubenswrapper[4888]: I1124 01:55:53.592673 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:55:53 crc kubenswrapper[4888]: I1124 01:55:53.592754 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:55:53 crc kubenswrapper[4888]: I1124 01:55:53.594242 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:55:53 crc kubenswrapper[4888]: I1124 01:55:53.594408 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23" gracePeriod=600 Nov 24 01:55:54 crc kubenswrapper[4888]: I1124 01:55:54.067312 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23" exitCode=0 Nov 24 01:55:54 crc kubenswrapper[4888]: I1124 01:55:54.067390 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23"} Nov 24 01:55:54 crc kubenswrapper[4888]: I1124 01:55:54.068039 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82"} Nov 24 01:55:54 crc kubenswrapper[4888]: I1124 01:55:54.068112 4888 scope.go:117] "RemoveContainer" containerID="df7bdb6d6c48db382282260a7be9d553829ccaa6713651f1417e5cb4cd4e6e3a" Nov 24 01:56:37 crc kubenswrapper[4888]: I1124 01:56:37.573228 4888 trace.go:236] Trace[1831205589]: "Calculate volume metrics of prometheus-metric-storage-db for pod openstack/prometheus-metric-storage-0" (24-Nov-2025 01:56:36.502) (total time: 1070ms): Nov 24 01:56:37 crc kubenswrapper[4888]: Trace[1831205589]: [1.070662644s] [1.070662644s] END Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.836654 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:57:43 crc kubenswrapper[4888]: E1124 01:57:43.840168 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="extract-content" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.840196 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="extract-content" Nov 24 01:57:43 crc kubenswrapper[4888]: E1124 01:57:43.840250 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="registry-server" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.840262 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="registry-server" Nov 24 01:57:43 crc kubenswrapper[4888]: E1124 01:57:43.840287 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="extract-utilities" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.840299 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="extract-utilities" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.840705 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="501b7734-df4d-4c4f-9bd0-56994ef6c447" containerName="registry-server" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.843338 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.847256 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.954767 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.954931 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r59b\" (UniqueName: \"kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:43 crc kubenswrapper[4888]: I1124 01:57:43.955210 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.057657 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.058110 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r59b\" (UniqueName: \"kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.058254 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.058276 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.058657 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.079782 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r59b\" (UniqueName: \"kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b\") pod \"redhat-operators-zrc5l\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.183789 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:44 crc kubenswrapper[4888]: I1124 01:57:44.720372 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:57:45 crc kubenswrapper[4888]: I1124 01:57:45.505129 4888 generic.go:334] "Generic (PLEG): container finished" podID="7e02c833-253f-462a-9e58-b9efd0630c60" containerID="2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6" exitCode=0 Nov 24 01:57:45 crc kubenswrapper[4888]: I1124 01:57:45.505181 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerDied","Data":"2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6"} Nov 24 01:57:45 crc kubenswrapper[4888]: I1124 01:57:45.506126 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerStarted","Data":"f085641c5234c0793b15c80faae6d7dfa1e5319a24b8577a04224bd1edd376bb"} Nov 24 01:57:47 crc kubenswrapper[4888]: I1124 01:57:47.530421 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerStarted","Data":"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874"} Nov 24 01:57:50 crc kubenswrapper[4888]: E1124 01:57:50.344129 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e02c833_253f_462a_9e58_b9efd0630c60.slice/crio-bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874.scope\": RecentStats: unable to find data in memory cache]" Nov 24 01:57:50 crc kubenswrapper[4888]: I1124 01:57:50.567844 4888 generic.go:334] "Generic (PLEG): container finished" podID="7e02c833-253f-462a-9e58-b9efd0630c60" containerID="bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874" exitCode=0 Nov 24 01:57:50 crc kubenswrapper[4888]: I1124 01:57:50.568188 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerDied","Data":"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874"} Nov 24 01:57:51 crc kubenswrapper[4888]: I1124 01:57:51.584407 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerStarted","Data":"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c"} Nov 24 01:57:51 crc kubenswrapper[4888]: I1124 01:57:51.614081 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zrc5l" podStartSLOduration=3.115139409 podStartE2EDuration="8.61405451s" podCreationTimestamp="2025-11-24 01:57:43 +0000 UTC" firstStartedPulling="2025-11-24 01:57:45.507097707 +0000 UTC m=+5568.089781751" lastFinishedPulling="2025-11-24 01:57:51.006012798 +0000 UTC m=+5573.588696852" observedRunningTime="2025-11-24 01:57:51.604090921 +0000 UTC m=+5574.186775005" watchObservedRunningTime="2025-11-24 01:57:51.61405451 +0000 UTC m=+5574.196738594" Nov 24 01:57:53 crc kubenswrapper[4888]: I1124 01:57:53.591747 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:57:53 crc kubenswrapper[4888]: I1124 01:57:53.592191 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:57:54 crc kubenswrapper[4888]: I1124 01:57:54.184645 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:54 crc kubenswrapper[4888]: I1124 01:57:54.185028 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:57:55 crc kubenswrapper[4888]: I1124 01:57:55.232433 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zrc5l" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" probeResult="failure" output=< Nov 24 01:57:55 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 01:57:55 crc kubenswrapper[4888]: > Nov 24 01:58:05 crc kubenswrapper[4888]: I1124 01:58:05.240336 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zrc5l" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" probeResult="failure" output=< Nov 24 01:58:05 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 01:58:05 crc kubenswrapper[4888]: > Nov 24 01:58:14 crc kubenswrapper[4888]: I1124 01:58:14.243464 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:58:14 crc kubenswrapper[4888]: I1124 01:58:14.313930 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:58:15 crc kubenswrapper[4888]: I1124 01:58:15.040150 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:58:15 crc kubenswrapper[4888]: I1124 01:58:15.871875 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zrc5l" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" containerID="cri-o://23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c" gracePeriod=2 Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.418940 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.454320 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r59b\" (UniqueName: \"kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b\") pod \"7e02c833-253f-462a-9e58-b9efd0630c60\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.454389 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content\") pod \"7e02c833-253f-462a-9e58-b9efd0630c60\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.454418 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities\") pod \"7e02c833-253f-462a-9e58-b9efd0630c60\" (UID: \"7e02c833-253f-462a-9e58-b9efd0630c60\") " Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.456051 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities" (OuterVolumeSpecName: "utilities") pod "7e02c833-253f-462a-9e58-b9efd0630c60" (UID: "7e02c833-253f-462a-9e58-b9efd0630c60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.461617 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b" (OuterVolumeSpecName: "kube-api-access-8r59b") pod "7e02c833-253f-462a-9e58-b9efd0630c60" (UID: "7e02c833-253f-462a-9e58-b9efd0630c60"). InnerVolumeSpecName "kube-api-access-8r59b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.557130 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e02c833-253f-462a-9e58-b9efd0630c60" (UID: "7e02c833-253f-462a-9e58-b9efd0630c60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.557755 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r59b\" (UniqueName: \"kubernetes.io/projected/7e02c833-253f-462a-9e58-b9efd0630c60-kube-api-access-8r59b\") on node \"crc\" DevicePath \"\"" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.557795 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.557829 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e02c833-253f-462a-9e58-b9efd0630c60-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.886030 4888 generic.go:334] "Generic (PLEG): container finished" podID="7e02c833-253f-462a-9e58-b9efd0630c60" containerID="23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c" exitCode=0 Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.886092 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerDied","Data":"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c"} Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.886134 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrc5l" event={"ID":"7e02c833-253f-462a-9e58-b9efd0630c60","Type":"ContainerDied","Data":"f085641c5234c0793b15c80faae6d7dfa1e5319a24b8577a04224bd1edd376bb"} Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.886158 4888 scope.go:117] "RemoveContainer" containerID="23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.886158 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrc5l" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.926587 4888 scope.go:117] "RemoveContainer" containerID="bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874" Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.933755 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.945534 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zrc5l"] Nov 24 01:58:16 crc kubenswrapper[4888]: I1124 01:58:16.958138 4888 scope.go:117] "RemoveContainer" containerID="2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.003690 4888 scope.go:117] "RemoveContainer" containerID="23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c" Nov 24 01:58:17 crc kubenswrapper[4888]: E1124 01:58:17.004564 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c\": container with ID starting with 23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c not found: ID does not exist" containerID="23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.004608 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c"} err="failed to get container status \"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c\": rpc error: code = NotFound desc = could not find container \"23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c\": container with ID starting with 23c60e5f85800e5b031e9093aaca7c14acb4c14346602ada2483486b4e34553c not found: ID does not exist" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.004636 4888 scope.go:117] "RemoveContainer" containerID="bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874" Nov 24 01:58:17 crc kubenswrapper[4888]: E1124 01:58:17.005181 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874\": container with ID starting with bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874 not found: ID does not exist" containerID="bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.005214 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874"} err="failed to get container status \"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874\": rpc error: code = NotFound desc = could not find container \"bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874\": container with ID starting with bf6498dad0da049943fa9a6f9b3082c10f97d139ea14691946e823961c174874 not found: ID does not exist" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.005236 4888 scope.go:117] "RemoveContainer" containerID="2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6" Nov 24 01:58:17 crc kubenswrapper[4888]: E1124 01:58:17.005603 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6\": container with ID starting with 2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6 not found: ID does not exist" containerID="2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6" Nov 24 01:58:17 crc kubenswrapper[4888]: I1124 01:58:17.005621 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6"} err="failed to get container status \"2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6\": rpc error: code = NotFound desc = could not find container \"2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6\": container with ID starting with 2129c008bad4e6e5267a19f6c9c39aeeb018ac91e36db50b069f873da20504e6 not found: ID does not exist" Nov 24 01:58:18 crc kubenswrapper[4888]: I1124 01:58:18.266777 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" path="/var/lib/kubelet/pods/7e02c833-253f-462a-9e58-b9efd0630c60/volumes" Nov 24 01:58:23 crc kubenswrapper[4888]: I1124 01:58:23.591196 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:58:23 crc kubenswrapper[4888]: I1124 01:58:23.591660 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:58:53 crc kubenswrapper[4888]: I1124 01:58:53.591914 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 01:58:53 crc kubenswrapper[4888]: I1124 01:58:53.592733 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 01:58:53 crc kubenswrapper[4888]: I1124 01:58:53.592855 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 01:58:53 crc kubenswrapper[4888]: I1124 01:58:53.594302 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 01:58:53 crc kubenswrapper[4888]: I1124 01:58:53.594433 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" gracePeriod=600 Nov 24 01:58:53 crc kubenswrapper[4888]: E1124 01:58:53.750180 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:58:54 crc kubenswrapper[4888]: I1124 01:58:54.367234 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" exitCode=0 Nov 24 01:58:54 crc kubenswrapper[4888]: I1124 01:58:54.367308 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82"} Nov 24 01:58:54 crc kubenswrapper[4888]: I1124 01:58:54.367357 4888 scope.go:117] "RemoveContainer" containerID="7ad12dab7410cc6a62d9d0dff7167c5a192a027e5b7b01a1372d735ad6ae3d23" Nov 24 01:58:54 crc kubenswrapper[4888]: I1124 01:58:54.368427 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:58:54 crc kubenswrapper[4888]: E1124 01:58:54.368915 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:59:06 crc kubenswrapper[4888]: I1124 01:59:06.247136 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:59:06 crc kubenswrapper[4888]: E1124 01:59:06.249116 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:59:20 crc kubenswrapper[4888]: I1124 01:59:20.246121 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:59:20 crc kubenswrapper[4888]: E1124 01:59:20.247013 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:59:32 crc kubenswrapper[4888]: I1124 01:59:32.245464 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:59:32 crc kubenswrapper[4888]: E1124 01:59:32.246444 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:59:45 crc kubenswrapper[4888]: I1124 01:59:45.245797 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:59:45 crc kubenswrapper[4888]: E1124 01:59:45.246617 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 01:59:56 crc kubenswrapper[4888]: I1124 01:59:56.249286 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 01:59:56 crc kubenswrapper[4888]: E1124 01:59:56.250303 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.149119 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n"] Nov 24 02:00:00 crc kubenswrapper[4888]: E1124 02:00:00.150031 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="extract-utilities" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.150045 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="extract-utilities" Nov 24 02:00:00 crc kubenswrapper[4888]: E1124 02:00:00.150094 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="extract-content" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.150100 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="extract-content" Nov 24 02:00:00 crc kubenswrapper[4888]: E1124 02:00:00.150115 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.150121 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.150324 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e02c833-253f-462a-9e58-b9efd0630c60" containerName="registry-server" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.151115 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.153196 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.153720 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.162376 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n"] Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.290742 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.290800 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.290945 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6n76\" (UniqueName: \"kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.392663 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.392980 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.393034 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6n76\" (UniqueName: \"kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.394973 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.399513 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.408966 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6n76\" (UniqueName: \"kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76\") pod \"collect-profiles-29399160-rj55n\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.487415 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:00 crc kubenswrapper[4888]: I1124 02:00:00.961678 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n"] Nov 24 02:00:00 crc kubenswrapper[4888]: W1124 02:00:00.970258 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod665c431f_1a41_40c7_a486_c18c47be6d72.slice/crio-4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647 WatchSource:0}: Error finding container 4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647: Status 404 returned error can't find the container with id 4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647 Nov 24 02:00:01 crc kubenswrapper[4888]: I1124 02:00:01.216480 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" event={"ID":"665c431f-1a41-40c7-a486-c18c47be6d72","Type":"ContainerStarted","Data":"b43b69fa6cad4c49cdb9666706072be630277ffeac37049dea6ee00bed2879cd"} Nov 24 02:00:01 crc kubenswrapper[4888]: I1124 02:00:01.216947 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" event={"ID":"665c431f-1a41-40c7-a486-c18c47be6d72","Type":"ContainerStarted","Data":"4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647"} Nov 24 02:00:01 crc kubenswrapper[4888]: I1124 02:00:01.235953 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" podStartSLOduration=1.235936881 podStartE2EDuration="1.235936881s" podCreationTimestamp="2025-11-24 02:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:00:01.231168637 +0000 UTC m=+5703.813852681" watchObservedRunningTime="2025-11-24 02:00:01.235936881 +0000 UTC m=+5703.818620925" Nov 24 02:00:02 crc kubenswrapper[4888]: I1124 02:00:02.231050 4888 generic.go:334] "Generic (PLEG): container finished" podID="665c431f-1a41-40c7-a486-c18c47be6d72" containerID="b43b69fa6cad4c49cdb9666706072be630277ffeac37049dea6ee00bed2879cd" exitCode=0 Nov 24 02:00:02 crc kubenswrapper[4888]: I1124 02:00:02.231241 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" event={"ID":"665c431f-1a41-40c7-a486-c18c47be6d72","Type":"ContainerDied","Data":"b43b69fa6cad4c49cdb9666706072be630277ffeac37049dea6ee00bed2879cd"} Nov 24 02:00:03 crc kubenswrapper[4888]: I1124 02:00:03.859514 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:03 crc kubenswrapper[4888]: I1124 02:00:03.989139 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume\") pod \"665c431f-1a41-40c7-a486-c18c47be6d72\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " Nov 24 02:00:03 crc kubenswrapper[4888]: I1124 02:00:03.989322 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6n76\" (UniqueName: \"kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76\") pod \"665c431f-1a41-40c7-a486-c18c47be6d72\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " Nov 24 02:00:03 crc kubenswrapper[4888]: I1124 02:00:03.989347 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume\") pod \"665c431f-1a41-40c7-a486-c18c47be6d72\" (UID: \"665c431f-1a41-40c7-a486-c18c47be6d72\") " Nov 24 02:00:03 crc kubenswrapper[4888]: I1124 02:00:03.989972 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume" (OuterVolumeSpecName: "config-volume") pod "665c431f-1a41-40c7-a486-c18c47be6d72" (UID: "665c431f-1a41-40c7-a486-c18c47be6d72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:03.998673 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "665c431f-1a41-40c7-a486-c18c47be6d72" (UID: "665c431f-1a41-40c7-a486-c18c47be6d72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:03.999413 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76" (OuterVolumeSpecName: "kube-api-access-g6n76") pod "665c431f-1a41-40c7-a486-c18c47be6d72" (UID: "665c431f-1a41-40c7-a486-c18c47be6d72"). InnerVolumeSpecName "kube-api-access-g6n76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.092654 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665c431f-1a41-40c7-a486-c18c47be6d72-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.092696 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6n76\" (UniqueName: \"kubernetes.io/projected/665c431f-1a41-40c7-a486-c18c47be6d72-kube-api-access-g6n76\") on node \"crc\" DevicePath \"\"" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.092709 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/665c431f-1a41-40c7-a486-c18c47be6d72-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.257663 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.265230 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n" event={"ID":"665c431f-1a41-40c7-a486-c18c47be6d72","Type":"ContainerDied","Data":"4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647"} Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.265281 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d9c07180921d45da7b68b6461f39376b6183e93b803282fc0adadc316b54647" Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.328066 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls"] Nov 24 02:00:04 crc kubenswrapper[4888]: I1124 02:00:04.336913 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399115-z9tls"] Nov 24 02:00:04 crc kubenswrapper[4888]: E1124 02:00:04.477290 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod665c431f_1a41_40c7_a486_c18c47be6d72.slice\": RecentStats: unable to find data in memory cache]" Nov 24 02:00:06 crc kubenswrapper[4888]: I1124 02:00:06.258289 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61ae0db9-fa05-4603-b9f8-7ff49fa36022" path="/var/lib/kubelet/pods/61ae0db9-fa05-4603-b9f8-7ff49fa36022/volumes" Nov 24 02:00:07 crc kubenswrapper[4888]: I1124 02:00:07.246605 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:00:07 crc kubenswrapper[4888]: E1124 02:00:07.247475 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:00:20 crc kubenswrapper[4888]: I1124 02:00:20.246602 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:00:20 crc kubenswrapper[4888]: E1124 02:00:20.248052 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:00:21 crc kubenswrapper[4888]: I1124 02:00:21.622371 4888 scope.go:117] "RemoveContainer" containerID="c6a5a5bb86a990bee92b608a77f9a4c69b77793503ad55fad09a4aabbb065f55" Nov 24 02:00:32 crc kubenswrapper[4888]: I1124 02:00:32.246319 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:00:32 crc kubenswrapper[4888]: E1124 02:00:32.247592 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:00:44 crc kubenswrapper[4888]: I1124 02:00:44.245699 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:00:44 crc kubenswrapper[4888]: E1124 02:00:44.246988 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:00:57 crc kubenswrapper[4888]: I1124 02:00:57.246183 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:00:57 crc kubenswrapper[4888]: E1124 02:00:57.247001 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.183736 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399161-vzqj4"] Nov 24 02:01:00 crc kubenswrapper[4888]: E1124 02:01:00.184672 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665c431f-1a41-40c7-a486-c18c47be6d72" containerName="collect-profiles" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.184687 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="665c431f-1a41-40c7-a486-c18c47be6d72" containerName="collect-profiles" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.184938 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="665c431f-1a41-40c7-a486-c18c47be6d72" containerName="collect-profiles" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.185795 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.197610 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399161-vzqj4"] Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.329033 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.329211 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.329296 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.329482 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrt2m\" (UniqueName: \"kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.431741 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.431903 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.432067 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrt2m\" (UniqueName: \"kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.432299 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.439310 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.439966 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.440627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.454788 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrt2m\" (UniqueName: \"kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m\") pod \"keystone-cron-29399161-vzqj4\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.511128 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:00 crc kubenswrapper[4888]: I1124 02:01:00.968364 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399161-vzqj4"] Nov 24 02:01:01 crc kubenswrapper[4888]: E1124 02:01:01.089544 4888 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.23:52120->38.102.83.23:41433: write tcp 38.102.83.23:52120->38.102.83.23:41433: write: broken pipe Nov 24 02:01:01 crc kubenswrapper[4888]: W1124 02:01:01.456201 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ab4f95d_24df_4a76_b678_a62601ae4375.slice/crio-571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16 WatchSource:0}: Error finding container 571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16: Status 404 returned error can't find the container with id 571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16 Nov 24 02:01:02 crc kubenswrapper[4888]: I1124 02:01:02.021955 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399161-vzqj4" event={"ID":"5ab4f95d-24df-4a76-b678-a62601ae4375","Type":"ContainerStarted","Data":"b65027d39a2883a8b96a072cb2bd473ab3d34613cd6845bdaa21338d3be181d4"} Nov 24 02:01:02 crc kubenswrapper[4888]: I1124 02:01:02.022286 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399161-vzqj4" event={"ID":"5ab4f95d-24df-4a76-b678-a62601ae4375","Type":"ContainerStarted","Data":"571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16"} Nov 24 02:01:02 crc kubenswrapper[4888]: I1124 02:01:02.038656 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399161-vzqj4" podStartSLOduration=2.038637087 podStartE2EDuration="2.038637087s" podCreationTimestamp="2025-11-24 02:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:01:02.03768688 +0000 UTC m=+5764.620370924" watchObservedRunningTime="2025-11-24 02:01:02.038637087 +0000 UTC m=+5764.621321131" Nov 24 02:01:05 crc kubenswrapper[4888]: I1124 02:01:05.060388 4888 generic.go:334] "Generic (PLEG): container finished" podID="5ab4f95d-24df-4a76-b678-a62601ae4375" containerID="b65027d39a2883a8b96a072cb2bd473ab3d34613cd6845bdaa21338d3be181d4" exitCode=0 Nov 24 02:01:05 crc kubenswrapper[4888]: I1124 02:01:05.060501 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399161-vzqj4" event={"ID":"5ab4f95d-24df-4a76-b678-a62601ae4375","Type":"ContainerDied","Data":"b65027d39a2883a8b96a072cb2bd473ab3d34613cd6845bdaa21338d3be181d4"} Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.581399 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.691642 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys\") pod \"5ab4f95d-24df-4a76-b678-a62601ae4375\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.691864 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data\") pod \"5ab4f95d-24df-4a76-b678-a62601ae4375\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.692146 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle\") pod \"5ab4f95d-24df-4a76-b678-a62601ae4375\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.692185 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrt2m\" (UniqueName: \"kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m\") pod \"5ab4f95d-24df-4a76-b678-a62601ae4375\" (UID: \"5ab4f95d-24df-4a76-b678-a62601ae4375\") " Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.698609 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m" (OuterVolumeSpecName: "kube-api-access-qrt2m") pod "5ab4f95d-24df-4a76-b678-a62601ae4375" (UID: "5ab4f95d-24df-4a76-b678-a62601ae4375"). InnerVolumeSpecName "kube-api-access-qrt2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.710756 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5ab4f95d-24df-4a76-b678-a62601ae4375" (UID: "5ab4f95d-24df-4a76-b678-a62601ae4375"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.722293 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ab4f95d-24df-4a76-b678-a62601ae4375" (UID: "5ab4f95d-24df-4a76-b678-a62601ae4375"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.779759 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data" (OuterVolumeSpecName: "config-data") pod "5ab4f95d-24df-4a76-b678-a62601ae4375" (UID: "5ab4f95d-24df-4a76-b678-a62601ae4375"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.795090 4888 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.795123 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrt2m\" (UniqueName: \"kubernetes.io/projected/5ab4f95d-24df-4a76-b678-a62601ae4375-kube-api-access-qrt2m\") on node \"crc\" DevicePath \"\"" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.795140 4888 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 02:01:06 crc kubenswrapper[4888]: I1124 02:01:06.795152 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab4f95d-24df-4a76-b678-a62601ae4375-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 02:01:07 crc kubenswrapper[4888]: I1124 02:01:07.092684 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399161-vzqj4" event={"ID":"5ab4f95d-24df-4a76-b678-a62601ae4375","Type":"ContainerDied","Data":"571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16"} Nov 24 02:01:07 crc kubenswrapper[4888]: I1124 02:01:07.093143 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="571baef6b3f6cda65fdbf1c053117289eb179b6a075dfe6ee7a542c52b110b16" Nov 24 02:01:07 crc kubenswrapper[4888]: I1124 02:01:07.092765 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399161-vzqj4" Nov 24 02:01:11 crc kubenswrapper[4888]: I1124 02:01:11.246115 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:01:11 crc kubenswrapper[4888]: E1124 02:01:11.247650 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:01:25 crc kubenswrapper[4888]: I1124 02:01:25.246238 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:01:25 crc kubenswrapper[4888]: E1124 02:01:25.247509 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:01:37 crc kubenswrapper[4888]: I1124 02:01:37.253888 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:01:37 crc kubenswrapper[4888]: E1124 02:01:37.255120 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.893548 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:01:50 crc kubenswrapper[4888]: E1124 02:01:50.895944 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab4f95d-24df-4a76-b678-a62601ae4375" containerName="keystone-cron" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.895974 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab4f95d-24df-4a76-b678-a62601ae4375" containerName="keystone-cron" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.896345 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab4f95d-24df-4a76-b678-a62601ae4375" containerName="keystone-cron" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.899733 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.928791 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.945066 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.945179 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:50 crc kubenswrapper[4888]: I1124 02:01:50.945266 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xmj\" (UniqueName: \"kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.047224 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xmj\" (UniqueName: \"kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.047391 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.047441 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.047946 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.048059 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.069414 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xmj\" (UniqueName: \"kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj\") pod \"certified-operators-rltk6\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.238544 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:01:51 crc kubenswrapper[4888]: I1124 02:01:51.762646 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:01:51 crc kubenswrapper[4888]: W1124 02:01:51.768992 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7af4a3f_5cf2_4e56_b2e8_8920da68d172.slice/crio-55723954702eb33345d21c114259a5e8106893249ff82761c2118a76cfd439c4 WatchSource:0}: Error finding container 55723954702eb33345d21c114259a5e8106893249ff82761c2118a76cfd439c4: Status 404 returned error can't find the container with id 55723954702eb33345d21c114259a5e8106893249ff82761c2118a76cfd439c4 Nov 24 02:01:52 crc kubenswrapper[4888]: I1124 02:01:52.245792 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:01:52 crc kubenswrapper[4888]: E1124 02:01:52.246497 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:01:52 crc kubenswrapper[4888]: I1124 02:01:52.716380 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerID="2c0d34bed156034be2bc30c4ef193345b8d7b159c799cefab448d71aba43ed31" exitCode=0 Nov 24 02:01:52 crc kubenswrapper[4888]: I1124 02:01:52.716456 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerDied","Data":"2c0d34bed156034be2bc30c4ef193345b8d7b159c799cefab448d71aba43ed31"} Nov 24 02:01:52 crc kubenswrapper[4888]: I1124 02:01:52.716727 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerStarted","Data":"55723954702eb33345d21c114259a5e8106893249ff82761c2118a76cfd439c4"} Nov 24 02:01:52 crc kubenswrapper[4888]: I1124 02:01:52.719564 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:01:53 crc kubenswrapper[4888]: I1124 02:01:53.730185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerStarted","Data":"b20155f86c62561efe46575e8f5f906c55c8ca6215dce8e18bf51b1d2f0ad2ee"} Nov 24 02:01:55 crc kubenswrapper[4888]: I1124 02:01:55.755299 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerID="b20155f86c62561efe46575e8f5f906c55c8ca6215dce8e18bf51b1d2f0ad2ee" exitCode=0 Nov 24 02:01:55 crc kubenswrapper[4888]: I1124 02:01:55.755426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerDied","Data":"b20155f86c62561efe46575e8f5f906c55c8ca6215dce8e18bf51b1d2f0ad2ee"} Nov 24 02:01:56 crc kubenswrapper[4888]: I1124 02:01:56.770074 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerStarted","Data":"83036c003db92e8331de40633f4ea5574c7bcf95ff4ddcc37455df05b238f32c"} Nov 24 02:01:56 crc kubenswrapper[4888]: I1124 02:01:56.815291 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rltk6" podStartSLOduration=3.370702673 podStartE2EDuration="6.815264023s" podCreationTimestamp="2025-11-24 02:01:50 +0000 UTC" firstStartedPulling="2025-11-24 02:01:52.719347539 +0000 UTC m=+5815.302031583" lastFinishedPulling="2025-11-24 02:01:56.163908849 +0000 UTC m=+5818.746592933" observedRunningTime="2025-11-24 02:01:56.802918227 +0000 UTC m=+5819.385602271" watchObservedRunningTime="2025-11-24 02:01:56.815264023 +0000 UTC m=+5819.397948107" Nov 24 02:02:01 crc kubenswrapper[4888]: I1124 02:02:01.239531 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:01 crc kubenswrapper[4888]: I1124 02:02:01.240122 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:01 crc kubenswrapper[4888]: I1124 02:02:01.907527 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:01 crc kubenswrapper[4888]: I1124 02:02:01.975768 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:02 crc kubenswrapper[4888]: I1124 02:02:02.144015 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:02:03 crc kubenswrapper[4888]: I1124 02:02:03.896190 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rltk6" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="registry-server" containerID="cri-o://83036c003db92e8331de40633f4ea5574c7bcf95ff4ddcc37455df05b238f32c" gracePeriod=2 Nov 24 02:02:04 crc kubenswrapper[4888]: I1124 02:02:04.913234 4888 generic.go:334] "Generic (PLEG): container finished" podID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerID="83036c003db92e8331de40633f4ea5574c7bcf95ff4ddcc37455df05b238f32c" exitCode=0 Nov 24 02:02:04 crc kubenswrapper[4888]: I1124 02:02:04.913300 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerDied","Data":"83036c003db92e8331de40633f4ea5574c7bcf95ff4ddcc37455df05b238f32c"} Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.019282 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.121441 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52xmj\" (UniqueName: \"kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj\") pod \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.121628 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities\") pod \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.121705 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content\") pod \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\" (UID: \"e7af4a3f-5cf2-4e56-b2e8-8920da68d172\") " Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.123436 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities" (OuterVolumeSpecName: "utilities") pod "e7af4a3f-5cf2-4e56-b2e8-8920da68d172" (UID: "e7af4a3f-5cf2-4e56-b2e8-8920da68d172"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.125969 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.142053 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj" (OuterVolumeSpecName: "kube-api-access-52xmj") pod "e7af4a3f-5cf2-4e56-b2e8-8920da68d172" (UID: "e7af4a3f-5cf2-4e56-b2e8-8920da68d172"). InnerVolumeSpecName "kube-api-access-52xmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.171044 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7af4a3f-5cf2-4e56-b2e8-8920da68d172" (UID: "e7af4a3f-5cf2-4e56-b2e8-8920da68d172"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.227187 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52xmj\" (UniqueName: \"kubernetes.io/projected/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-kube-api-access-52xmj\") on node \"crc\" DevicePath \"\"" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.227222 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7af4a3f-5cf2-4e56-b2e8-8920da68d172-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.245061 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:02:05 crc kubenswrapper[4888]: E1124 02:02:05.245347 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.932460 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rltk6" event={"ID":"e7af4a3f-5cf2-4e56-b2e8-8920da68d172","Type":"ContainerDied","Data":"55723954702eb33345d21c114259a5e8106893249ff82761c2118a76cfd439c4"} Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.932761 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rltk6" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.933014 4888 scope.go:117] "RemoveContainer" containerID="83036c003db92e8331de40633f4ea5574c7bcf95ff4ddcc37455df05b238f32c" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.971974 4888 scope.go:117] "RemoveContainer" containerID="b20155f86c62561efe46575e8f5f906c55c8ca6215dce8e18bf51b1d2f0ad2ee" Nov 24 02:02:05 crc kubenswrapper[4888]: I1124 02:02:05.999669 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:02:06 crc kubenswrapper[4888]: I1124 02:02:06.014657 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rltk6"] Nov 24 02:02:06 crc kubenswrapper[4888]: I1124 02:02:06.025639 4888 scope.go:117] "RemoveContainer" containerID="2c0d34bed156034be2bc30c4ef193345b8d7b159c799cefab448d71aba43ed31" Nov 24 02:02:06 crc kubenswrapper[4888]: I1124 02:02:06.257533 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" path="/var/lib/kubelet/pods/e7af4a3f-5cf2-4e56-b2e8-8920da68d172/volumes" Nov 24 02:02:18 crc kubenswrapper[4888]: I1124 02:02:18.270852 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:02:18 crc kubenswrapper[4888]: E1124 02:02:18.272004 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:02:31 crc kubenswrapper[4888]: I1124 02:02:31.245441 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:02:31 crc kubenswrapper[4888]: E1124 02:02:31.246355 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:02:42 crc kubenswrapper[4888]: I1124 02:02:42.246914 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:02:42 crc kubenswrapper[4888]: E1124 02:02:42.248565 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:02:56 crc kubenswrapper[4888]: I1124 02:02:56.245848 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:02:56 crc kubenswrapper[4888]: E1124 02:02:56.246496 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:03:11 crc kubenswrapper[4888]: I1124 02:03:11.246331 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:03:11 crc kubenswrapper[4888]: E1124 02:03:11.247701 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:03:24 crc kubenswrapper[4888]: I1124 02:03:24.245349 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:03:24 crc kubenswrapper[4888]: E1124 02:03:24.247446 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:03:35 crc kubenswrapper[4888]: I1124 02:03:35.244832 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:03:35 crc kubenswrapper[4888]: E1124 02:03:35.245470 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:03:49 crc kubenswrapper[4888]: I1124 02:03:49.245753 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:03:49 crc kubenswrapper[4888]: E1124 02:03:49.247022 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:04:02 crc kubenswrapper[4888]: I1124 02:04:02.246751 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:04:03 crc kubenswrapper[4888]: I1124 02:04:03.328580 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6"} Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.267724 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:26 crc kubenswrapper[4888]: E1124 02:04:26.268750 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="registry-server" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.268767 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="registry-server" Nov 24 02:04:26 crc kubenswrapper[4888]: E1124 02:04:26.268783 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="extract-content" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.268790 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="extract-content" Nov 24 02:04:26 crc kubenswrapper[4888]: E1124 02:04:26.268831 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="extract-utilities" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.268840 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="extract-utilities" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.269138 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7af4a3f-5cf2-4e56-b2e8-8920da68d172" containerName="registry-server" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.273554 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.273709 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.409087 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.409664 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.409714 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmbbm\" (UniqueName: \"kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.511626 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.511689 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.512262 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.512285 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.512359 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmbbm\" (UniqueName: \"kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.534137 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmbbm\" (UniqueName: \"kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm\") pod \"redhat-marketplace-tw5l8\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:26 crc kubenswrapper[4888]: I1124 02:04:26.613125 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.117252 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.672313 4888 generic.go:334] "Generic (PLEG): container finished" podID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerID="62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59" exitCode=0 Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.672390 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerDied","Data":"62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59"} Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.672718 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerStarted","Data":"3cc8ac47c13f1e1eb52ad01ed46f2d996aaaae319b029560b553f641b07f53e8"} Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.881749 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.885015 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.893749 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gc5cj" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.894050 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.894194 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.894428 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.897638 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948144 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948436 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-956zg\" (UniqueName: \"kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948552 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948648 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948758 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.948917 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.949061 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.949256 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:27 crc kubenswrapper[4888]: I1124 02:04:27.949329 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052204 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052458 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052606 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052638 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052667 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052762 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052799 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-956zg\" (UniqueName: \"kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052885 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.052959 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.053483 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.053768 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.053842 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.056408 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.058886 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.059274 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.059893 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.060552 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.071592 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-956zg\" (UniqueName: \"kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.095122 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.217364 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 02:04:28 crc kubenswrapper[4888]: I1124 02:04:28.711280 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 02:04:29 crc kubenswrapper[4888]: I1124 02:04:29.693970 4888 generic.go:334] "Generic (PLEG): container finished" podID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerID="dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef" exitCode=0 Nov 24 02:04:29 crc kubenswrapper[4888]: I1124 02:04:29.694147 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerDied","Data":"dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef"} Nov 24 02:04:29 crc kubenswrapper[4888]: I1124 02:04:29.696906 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cc931795-8f17-4afd-a16c-264855b1755c","Type":"ContainerStarted","Data":"af22224b967ee37b219f01467a10ceaccbbad177e240d97c7d0bb474ae4d8e2e"} Nov 24 02:04:31 crc kubenswrapper[4888]: I1124 02:04:31.724027 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerStarted","Data":"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276"} Nov 24 02:04:31 crc kubenswrapper[4888]: I1124 02:04:31.755930 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tw5l8" podStartSLOduration=2.8989777759999997 podStartE2EDuration="5.755913721s" podCreationTimestamp="2025-11-24 02:04:26 +0000 UTC" firstStartedPulling="2025-11-24 02:04:27.674509325 +0000 UTC m=+5970.257193389" lastFinishedPulling="2025-11-24 02:04:30.53144529 +0000 UTC m=+5973.114129334" observedRunningTime="2025-11-24 02:04:31.748090812 +0000 UTC m=+5974.330774866" watchObservedRunningTime="2025-11-24 02:04:31.755913721 +0000 UTC m=+5974.338597765" Nov 24 02:04:36 crc kubenswrapper[4888]: I1124 02:04:36.613955 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:36 crc kubenswrapper[4888]: I1124 02:04:36.614858 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:36 crc kubenswrapper[4888]: I1124 02:04:36.678234 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:36 crc kubenswrapper[4888]: I1124 02:04:36.858994 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:37 crc kubenswrapper[4888]: I1124 02:04:37.445566 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:38 crc kubenswrapper[4888]: I1124 02:04:38.827061 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tw5l8" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="registry-server" containerID="cri-o://4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276" gracePeriod=2 Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.637666 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.750599 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content\") pod \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.750784 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmbbm\" (UniqueName: \"kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm\") pod \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.751072 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities\") pod \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\" (UID: \"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed\") " Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.751701 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities" (OuterVolumeSpecName: "utilities") pod "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" (UID: "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.761422 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm" (OuterVolumeSpecName: "kube-api-access-mmbbm") pod "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" (UID: "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed"). InnerVolumeSpecName "kube-api-access-mmbbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.767064 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" (UID: "09f4cd6a-6016-49d4-97d6-1bffe2dc0bed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.842379 4888 generic.go:334] "Generic (PLEG): container finished" podID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerID="4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276" exitCode=0 Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.842420 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerDied","Data":"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276"} Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.842481 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tw5l8" event={"ID":"09f4cd6a-6016-49d4-97d6-1bffe2dc0bed","Type":"ContainerDied","Data":"3cc8ac47c13f1e1eb52ad01ed46f2d996aaaae319b029560b553f641b07f53e8"} Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.842502 4888 scope.go:117] "RemoveContainer" containerID="4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.842649 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tw5l8" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.854595 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.854627 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.854641 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmbbm\" (UniqueName: \"kubernetes.io/projected/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed-kube-api-access-mmbbm\") on node \"crc\" DevicePath \"\"" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.879864 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.880915 4888 scope.go:117] "RemoveContainer" containerID="dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.888267 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tw5l8"] Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.910472 4888 scope.go:117] "RemoveContainer" containerID="62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.958468 4888 scope.go:117] "RemoveContainer" containerID="4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276" Nov 24 02:04:39 crc kubenswrapper[4888]: E1124 02:04:39.959052 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276\": container with ID starting with 4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276 not found: ID does not exist" containerID="4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.959084 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276"} err="failed to get container status \"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276\": rpc error: code = NotFound desc = could not find container \"4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276\": container with ID starting with 4ae448f4ff977bd44e41ee3340fad6ae4d1be50a03e7aa9acfe7cb513fa93276 not found: ID does not exist" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.959104 4888 scope.go:117] "RemoveContainer" containerID="dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef" Nov 24 02:04:39 crc kubenswrapper[4888]: E1124 02:04:39.959443 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef\": container with ID starting with dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef not found: ID does not exist" containerID="dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.959488 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef"} err="failed to get container status \"dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef\": rpc error: code = NotFound desc = could not find container \"dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef\": container with ID starting with dcac6144cf102518658c6bdb24c0b911f2c6974a2cdfeab4ede25dec73377cef not found: ID does not exist" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.959518 4888 scope.go:117] "RemoveContainer" containerID="62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59" Nov 24 02:04:39 crc kubenswrapper[4888]: E1124 02:04:39.959902 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59\": container with ID starting with 62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59 not found: ID does not exist" containerID="62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59" Nov 24 02:04:39 crc kubenswrapper[4888]: I1124 02:04:39.959958 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59"} err="failed to get container status \"62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59\": rpc error: code = NotFound desc = could not find container \"62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59\": container with ID starting with 62bd419a6ceff3109ed2aa5dfad3ad255fcfdcad589032839f33fc3cc6d39c59 not found: ID does not exist" Nov 24 02:04:40 crc kubenswrapper[4888]: I1124 02:04:40.256574 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" path="/var/lib/kubelet/pods/09f4cd6a-6016-49d4-97d6-1bffe2dc0bed/volumes" Nov 24 02:04:58 crc kubenswrapper[4888]: E1124 02:04:58.756085 4888 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 02:04:58 crc kubenswrapper[4888]: E1124 02:04:58.760709 4888 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-956zg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(cc931795-8f17-4afd-a16c-264855b1755c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 02:04:58 crc kubenswrapper[4888]: E1124 02:04:58.762020 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="cc931795-8f17-4afd-a16c-264855b1755c" Nov 24 02:04:59 crc kubenswrapper[4888]: E1124 02:04:59.063096 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="cc931795-8f17-4afd-a16c-264855b1755c" Nov 24 02:05:13 crc kubenswrapper[4888]: I1124 02:05:13.697569 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 02:05:15 crc kubenswrapper[4888]: I1124 02:05:15.277244 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cc931795-8f17-4afd-a16c-264855b1755c","Type":"ContainerStarted","Data":"10ceea90344fa48a36b4bc996c2a9c6e0ad225c08f50143e81749208083fb023"} Nov 24 02:05:15 crc kubenswrapper[4888]: I1124 02:05:15.306491 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.335189471 podStartE2EDuration="49.306467072s" podCreationTimestamp="2025-11-24 02:04:26 +0000 UTC" firstStartedPulling="2025-11-24 02:04:28.718974372 +0000 UTC m=+5971.301658416" lastFinishedPulling="2025-11-24 02:05:13.690251943 +0000 UTC m=+6016.272936017" observedRunningTime="2025-11-24 02:05:15.296920954 +0000 UTC m=+6017.879605018" watchObservedRunningTime="2025-11-24 02:05:15.306467072 +0000 UTC m=+6017.889151126" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.199383 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:10 crc kubenswrapper[4888]: E1124 02:06:10.200325 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="registry-server" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.200338 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="registry-server" Nov 24 02:06:10 crc kubenswrapper[4888]: E1124 02:06:10.200380 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="extract-utilities" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.200387 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="extract-utilities" Nov 24 02:06:10 crc kubenswrapper[4888]: E1124 02:06:10.200407 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="extract-content" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.200413 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="extract-content" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.200612 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f4cd6a-6016-49d4-97d6-1bffe2dc0bed" containerName="registry-server" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.202144 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.223393 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.341781 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.341861 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvb7k\" (UniqueName: \"kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.342184 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.444542 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.444698 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.444727 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvb7k\" (UniqueName: \"kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.447950 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.448017 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.469589 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvb7k\" (UniqueName: \"kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k\") pod \"community-operators-ppz98\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:10 crc kubenswrapper[4888]: I1124 02:06:10.526539 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:11 crc kubenswrapper[4888]: I1124 02:06:11.633455 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:12 crc kubenswrapper[4888]: I1124 02:06:12.035734 4888 generic.go:334] "Generic (PLEG): container finished" podID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerID="3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683" exitCode=0 Nov 24 02:06:12 crc kubenswrapper[4888]: I1124 02:06:12.035888 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerDied","Data":"3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683"} Nov 24 02:06:12 crc kubenswrapper[4888]: I1124 02:06:12.036094 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerStarted","Data":"ebc430d6bc1357f33d5f979ba2beb916d3ee88338d1ff130377d12dd08445972"} Nov 24 02:06:13 crc kubenswrapper[4888]: I1124 02:06:13.048873 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerStarted","Data":"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f"} Nov 24 02:06:15 crc kubenswrapper[4888]: I1124 02:06:15.077045 4888 generic.go:334] "Generic (PLEG): container finished" podID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerID="ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f" exitCode=0 Nov 24 02:06:15 crc kubenswrapper[4888]: I1124 02:06:15.077490 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerDied","Data":"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f"} Nov 24 02:06:16 crc kubenswrapper[4888]: I1124 02:06:16.089646 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerStarted","Data":"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e"} Nov 24 02:06:16 crc kubenswrapper[4888]: I1124 02:06:16.120413 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ppz98" podStartSLOduration=2.543467224 podStartE2EDuration="6.120385259s" podCreationTimestamp="2025-11-24 02:06:10 +0000 UTC" firstStartedPulling="2025-11-24 02:06:12.037304987 +0000 UTC m=+6074.619989031" lastFinishedPulling="2025-11-24 02:06:15.614222992 +0000 UTC m=+6078.196907066" observedRunningTime="2025-11-24 02:06:16.111993604 +0000 UTC m=+6078.694677648" watchObservedRunningTime="2025-11-24 02:06:16.120385259 +0000 UTC m=+6078.703069313" Nov 24 02:06:20 crc kubenswrapper[4888]: I1124 02:06:20.526739 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:20 crc kubenswrapper[4888]: I1124 02:06:20.527394 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:20 crc kubenswrapper[4888]: I1124 02:06:20.581474 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:21 crc kubenswrapper[4888]: I1124 02:06:21.217180 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:21 crc kubenswrapper[4888]: I1124 02:06:21.270477 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:23 crc kubenswrapper[4888]: I1124 02:06:23.165768 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ppz98" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="registry-server" containerID="cri-o://007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e" gracePeriod=2 Nov 24 02:06:23 crc kubenswrapper[4888]: I1124 02:06:23.591358 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:06:23 crc kubenswrapper[4888]: I1124 02:06:23.596085 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:06:23 crc kubenswrapper[4888]: I1124 02:06:23.959356 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.070480 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content\") pod \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.070991 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvb7k\" (UniqueName: \"kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k\") pod \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.071203 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities\") pod \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\" (UID: \"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5\") " Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.072011 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities" (OuterVolumeSpecName: "utilities") pod "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" (UID: "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.084980 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k" (OuterVolumeSpecName: "kube-api-access-fvb7k") pod "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" (UID: "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5"). InnerVolumeSpecName "kube-api-access-fvb7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.125538 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" (UID: "11fe719e-4d39-4e4e-a7d5-98c53c29fdd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.174370 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.174429 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvb7k\" (UniqueName: \"kubernetes.io/projected/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-kube-api-access-fvb7k\") on node \"crc\" DevicePath \"\"" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.174443 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.179566 4888 generic.go:334] "Generic (PLEG): container finished" podID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerID="007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e" exitCode=0 Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.179610 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerDied","Data":"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e"} Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.179639 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppz98" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.179674 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppz98" event={"ID":"11fe719e-4d39-4e4e-a7d5-98c53c29fdd5","Type":"ContainerDied","Data":"ebc430d6bc1357f33d5f979ba2beb916d3ee88338d1ff130377d12dd08445972"} Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.179694 4888 scope.go:117] "RemoveContainer" containerID="007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.212727 4888 scope.go:117] "RemoveContainer" containerID="ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.224161 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.238315 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ppz98"] Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.258099 4888 scope.go:117] "RemoveContainer" containerID="3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.264284 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" path="/var/lib/kubelet/pods/11fe719e-4d39-4e4e-a7d5-98c53c29fdd5/volumes" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.289650 4888 scope.go:117] "RemoveContainer" containerID="007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e" Nov 24 02:06:24 crc kubenswrapper[4888]: E1124 02:06:24.290751 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e\": container with ID starting with 007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e not found: ID does not exist" containerID="007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.290820 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e"} err="failed to get container status \"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e\": rpc error: code = NotFound desc = could not find container \"007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e\": container with ID starting with 007ea20ca947a6b370dc8789d8a870496f2b1f5091f3db3999483269a603312e not found: ID does not exist" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.290855 4888 scope.go:117] "RemoveContainer" containerID="ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f" Nov 24 02:06:24 crc kubenswrapper[4888]: E1124 02:06:24.291303 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f\": container with ID starting with ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f not found: ID does not exist" containerID="ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.291346 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f"} err="failed to get container status \"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f\": rpc error: code = NotFound desc = could not find container \"ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f\": container with ID starting with ff5d4f80b037c86cdb0339b55f16ee0171434476ff15783064af989a71145e4f not found: ID does not exist" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.291375 4888 scope.go:117] "RemoveContainer" containerID="3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683" Nov 24 02:06:24 crc kubenswrapper[4888]: E1124 02:06:24.292151 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683\": container with ID starting with 3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683 not found: ID does not exist" containerID="3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683" Nov 24 02:06:24 crc kubenswrapper[4888]: I1124 02:06:24.292181 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683"} err="failed to get container status \"3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683\": rpc error: code = NotFound desc = could not find container \"3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683\": container with ID starting with 3ec93cf3f6aa530b51acab89431a113c989f366bcd83feded4ebf703a8b17683 not found: ID does not exist" Nov 24 02:06:53 crc kubenswrapper[4888]: I1124 02:06:53.592892 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:06:53 crc kubenswrapper[4888]: I1124 02:06:53.596332 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.592199 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.592886 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.592930 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.593800 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.593871 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6" gracePeriod=600 Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.796952 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6" exitCode=0 Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.797020 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6"} Nov 24 02:07:23 crc kubenswrapper[4888]: I1124 02:07:23.797314 4888 scope.go:117] "RemoveContainer" containerID="fd6f5e83f19127bb8800f58627492c555dc3c2277beeeea62097c70db4470e82" Nov 24 02:07:24 crc kubenswrapper[4888]: I1124 02:07:24.812585 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f"} Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.150236 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:14 crc kubenswrapper[4888]: E1124 02:08:14.155581 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="registry-server" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.155680 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="registry-server" Nov 24 02:08:14 crc kubenswrapper[4888]: E1124 02:08:14.155748 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="extract-content" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.155807 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="extract-content" Nov 24 02:08:14 crc kubenswrapper[4888]: E1124 02:08:14.155905 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="extract-utilities" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.155962 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="extract-utilities" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.156246 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fe719e-4d39-4e4e-a7d5-98c53c29fdd5" containerName="registry-server" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.157861 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.183031 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.225562 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.225685 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktwrz\" (UniqueName: \"kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.225759 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.327787 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktwrz\" (UniqueName: \"kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.328236 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.328568 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.330467 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.330519 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.350420 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktwrz\" (UniqueName: \"kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz\") pod \"redhat-operators-mjzpr\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.480511 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:14 crc kubenswrapper[4888]: I1124 02:08:14.957757 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:14 crc kubenswrapper[4888]: W1124 02:08:14.963568 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52ab1ff_1c04_4b74_9064_7643b61f1329.slice/crio-c50f1d43ddbe92084cc6bf9e314979eb35673b845d646eab7a47f1643d8ccf9c WatchSource:0}: Error finding container c50f1d43ddbe92084cc6bf9e314979eb35673b845d646eab7a47f1643d8ccf9c: Status 404 returned error can't find the container with id c50f1d43ddbe92084cc6bf9e314979eb35673b845d646eab7a47f1643d8ccf9c Nov 24 02:08:15 crc kubenswrapper[4888]: I1124 02:08:15.415924 4888 generic.go:334] "Generic (PLEG): container finished" podID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerID="737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377" exitCode=0 Nov 24 02:08:15 crc kubenswrapper[4888]: I1124 02:08:15.415989 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerDied","Data":"737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377"} Nov 24 02:08:15 crc kubenswrapper[4888]: I1124 02:08:15.416163 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerStarted","Data":"c50f1d43ddbe92084cc6bf9e314979eb35673b845d646eab7a47f1643d8ccf9c"} Nov 24 02:08:15 crc kubenswrapper[4888]: I1124 02:08:15.419149 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:08:17 crc kubenswrapper[4888]: I1124 02:08:17.436966 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerStarted","Data":"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99"} Nov 24 02:08:20 crc kubenswrapper[4888]: I1124 02:08:20.468248 4888 generic.go:334] "Generic (PLEG): container finished" podID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerID="5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99" exitCode=0 Nov 24 02:08:20 crc kubenswrapper[4888]: I1124 02:08:20.468332 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerDied","Data":"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99"} Nov 24 02:08:21 crc kubenswrapper[4888]: I1124 02:08:21.482426 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerStarted","Data":"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4"} Nov 24 02:08:21 crc kubenswrapper[4888]: I1124 02:08:21.512961 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mjzpr" podStartSLOduration=1.972115977 podStartE2EDuration="7.51294064s" podCreationTimestamp="2025-11-24 02:08:14 +0000 UTC" firstStartedPulling="2025-11-24 02:08:15.418156398 +0000 UTC m=+6198.000840432" lastFinishedPulling="2025-11-24 02:08:20.958981041 +0000 UTC m=+6203.541665095" observedRunningTime="2025-11-24 02:08:21.509752391 +0000 UTC m=+6204.092436475" watchObservedRunningTime="2025-11-24 02:08:21.51294064 +0000 UTC m=+6204.095624684" Nov 24 02:08:24 crc kubenswrapper[4888]: I1124 02:08:24.481450 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:24 crc kubenswrapper[4888]: I1124 02:08:24.482178 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:25 crc kubenswrapper[4888]: I1124 02:08:25.561923 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mjzpr" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" probeResult="failure" output=< Nov 24 02:08:25 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:08:25 crc kubenswrapper[4888]: > Nov 24 02:08:35 crc kubenswrapper[4888]: I1124 02:08:35.531156 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mjzpr" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" probeResult="failure" output=< Nov 24 02:08:35 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:08:35 crc kubenswrapper[4888]: > Nov 24 02:08:45 crc kubenswrapper[4888]: I1124 02:08:45.544444 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mjzpr" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" probeResult="failure" output=< Nov 24 02:08:45 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:08:45 crc kubenswrapper[4888]: > Nov 24 02:08:54 crc kubenswrapper[4888]: I1124 02:08:54.570221 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:54 crc kubenswrapper[4888]: I1124 02:08:54.637164 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:54 crc kubenswrapper[4888]: I1124 02:08:54.825145 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:55 crc kubenswrapper[4888]: I1124 02:08:55.885091 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mjzpr" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" containerID="cri-o://44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4" gracePeriod=2 Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.544098 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.743647 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content\") pod \"e52ab1ff-1c04-4b74-9064-7643b61f1329\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.743832 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities\") pod \"e52ab1ff-1c04-4b74-9064-7643b61f1329\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.744057 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktwrz\" (UniqueName: \"kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz\") pod \"e52ab1ff-1c04-4b74-9064-7643b61f1329\" (UID: \"e52ab1ff-1c04-4b74-9064-7643b61f1329\") " Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.745442 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities" (OuterVolumeSpecName: "utilities") pod "e52ab1ff-1c04-4b74-9064-7643b61f1329" (UID: "e52ab1ff-1c04-4b74-9064-7643b61f1329"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.769107 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz" (OuterVolumeSpecName: "kube-api-access-ktwrz") pod "e52ab1ff-1c04-4b74-9064-7643b61f1329" (UID: "e52ab1ff-1c04-4b74-9064-7643b61f1329"). InnerVolumeSpecName "kube-api-access-ktwrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.847709 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktwrz\" (UniqueName: \"kubernetes.io/projected/e52ab1ff-1c04-4b74-9064-7643b61f1329-kube-api-access-ktwrz\") on node \"crc\" DevicePath \"\"" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.847741 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.895266 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e52ab1ff-1c04-4b74-9064-7643b61f1329" (UID: "e52ab1ff-1c04-4b74-9064-7643b61f1329"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.898915 4888 generic.go:334] "Generic (PLEG): container finished" podID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerID="44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4" exitCode=0 Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.898957 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerDied","Data":"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4"} Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.898985 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjzpr" event={"ID":"e52ab1ff-1c04-4b74-9064-7643b61f1329","Type":"ContainerDied","Data":"c50f1d43ddbe92084cc6bf9e314979eb35673b845d646eab7a47f1643d8ccf9c"} Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.899005 4888 scope.go:117] "RemoveContainer" containerID="44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.899190 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjzpr" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.930988 4888 scope.go:117] "RemoveContainer" containerID="5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.936250 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.944974 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mjzpr"] Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.949529 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52ab1ff-1c04-4b74-9064-7643b61f1329-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:08:56 crc kubenswrapper[4888]: I1124 02:08:56.967337 4888 scope.go:117] "RemoveContainer" containerID="737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.011781 4888 scope.go:117] "RemoveContainer" containerID="44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4" Nov 24 02:08:57 crc kubenswrapper[4888]: E1124 02:08:57.012344 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4\": container with ID starting with 44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4 not found: ID does not exist" containerID="44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.012380 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4"} err="failed to get container status \"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4\": rpc error: code = NotFound desc = could not find container \"44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4\": container with ID starting with 44955d782bd9b745d57b712e6a85badd3c584490c84e3c92ffd8c769b88ed3d4 not found: ID does not exist" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.012407 4888 scope.go:117] "RemoveContainer" containerID="5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99" Nov 24 02:08:57 crc kubenswrapper[4888]: E1124 02:08:57.012716 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99\": container with ID starting with 5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99 not found: ID does not exist" containerID="5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.012740 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99"} err="failed to get container status \"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99\": rpc error: code = NotFound desc = could not find container \"5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99\": container with ID starting with 5b66f465aa34c64c6813a538184f05164baf4d0ef62fa264d8e6c5655a0f8f99 not found: ID does not exist" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.012757 4888 scope.go:117] "RemoveContainer" containerID="737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377" Nov 24 02:08:57 crc kubenswrapper[4888]: E1124 02:08:57.013139 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377\": container with ID starting with 737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377 not found: ID does not exist" containerID="737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377" Nov 24 02:08:57 crc kubenswrapper[4888]: I1124 02:08:57.013184 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377"} err="failed to get container status \"737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377\": rpc error: code = NotFound desc = could not find container \"737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377\": container with ID starting with 737a936e3d1ca0aa5a92adb264cfdb2eb1fa63f7e6def00be6385df05b555377 not found: ID does not exist" Nov 24 02:08:58 crc kubenswrapper[4888]: I1124 02:08:58.265914 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" path="/var/lib/kubelet/pods/e52ab1ff-1c04-4b74-9064-7643b61f1329/volumes" Nov 24 02:09:23 crc kubenswrapper[4888]: I1124 02:09:23.591424 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:09:23 crc kubenswrapper[4888]: I1124 02:09:23.592341 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:09:53 crc kubenswrapper[4888]: I1124 02:09:53.591487 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:09:53 crc kubenswrapper[4888]: I1124 02:09:53.593595 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.591953 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.592668 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.592751 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.593669 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.593744 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" gracePeriod=600 Nov 24 02:10:23 crc kubenswrapper[4888]: E1124 02:10:23.714484 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.932752 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" exitCode=0 Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.932800 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f"} Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.932855 4888 scope.go:117] "RemoveContainer" containerID="707663d3f38c91a6336780ab764a1821e78ec51cf13e9f3689d781aafceb79b6" Nov 24 02:10:23 crc kubenswrapper[4888]: I1124 02:10:23.933515 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:10:23 crc kubenswrapper[4888]: E1124 02:10:23.933856 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:10:39 crc kubenswrapper[4888]: I1124 02:10:39.245783 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:10:39 crc kubenswrapper[4888]: E1124 02:10:39.246868 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:10:51 crc kubenswrapper[4888]: I1124 02:10:51.246130 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:10:51 crc kubenswrapper[4888]: E1124 02:10:51.247398 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:11:05 crc kubenswrapper[4888]: I1124 02:11:05.246125 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:11:05 crc kubenswrapper[4888]: E1124 02:11:05.247489 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:11:19 crc kubenswrapper[4888]: I1124 02:11:19.245936 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:11:19 crc kubenswrapper[4888]: E1124 02:11:19.248443 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:11:34 crc kubenswrapper[4888]: I1124 02:11:34.246019 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:11:34 crc kubenswrapper[4888]: E1124 02:11:34.246793 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:11:49 crc kubenswrapper[4888]: I1124 02:11:49.245658 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:11:49 crc kubenswrapper[4888]: E1124 02:11:49.246790 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:12:01 crc kubenswrapper[4888]: I1124 02:12:01.245650 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:12:01 crc kubenswrapper[4888]: E1124 02:12:01.246640 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:12:12 crc kubenswrapper[4888]: I1124 02:12:12.246284 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:12:12 crc kubenswrapper[4888]: E1124 02:12:12.247382 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:12:26 crc kubenswrapper[4888]: I1124 02:12:26.245074 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:12:26 crc kubenswrapper[4888]: E1124 02:12:26.245893 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:12:40 crc kubenswrapper[4888]: I1124 02:12:40.246436 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:12:40 crc kubenswrapper[4888]: E1124 02:12:40.247397 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:12:54 crc kubenswrapper[4888]: I1124 02:12:54.246297 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:12:54 crc kubenswrapper[4888]: E1124 02:12:54.247129 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.317112 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:00 crc kubenswrapper[4888]: E1124 02:13:00.318684 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="extract-content" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.318715 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="extract-content" Nov 24 02:13:00 crc kubenswrapper[4888]: E1124 02:13:00.318745 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="extract-utilities" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.318760 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="extract-utilities" Nov 24 02:13:00 crc kubenswrapper[4888]: E1124 02:13:00.318801 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.318848 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.319284 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52ab1ff-1c04-4b74-9064-7643b61f1329" containerName="registry-server" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.323502 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.348477 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.397623 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.398288 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.398440 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgt2k\" (UniqueName: \"kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.501457 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.501518 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.501595 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgt2k\" (UniqueName: \"kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.502190 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.502776 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.523320 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgt2k\" (UniqueName: \"kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k\") pod \"certified-operators-r4w94\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:00 crc kubenswrapper[4888]: I1124 02:13:00.662450 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:01 crc kubenswrapper[4888]: I1124 02:13:01.135465 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:01 crc kubenswrapper[4888]: I1124 02:13:01.865358 4888 generic.go:334] "Generic (PLEG): container finished" podID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerID="6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e" exitCode=0 Nov 24 02:13:01 crc kubenswrapper[4888]: I1124 02:13:01.865482 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerDied","Data":"6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e"} Nov 24 02:13:01 crc kubenswrapper[4888]: I1124 02:13:01.865798 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerStarted","Data":"f39fad44bb827202f97564c2a258402fa4b0c9cf3a22396b2c23dfcb311e039c"} Nov 24 02:13:02 crc kubenswrapper[4888]: I1124 02:13:02.883103 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerStarted","Data":"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a"} Nov 24 02:13:03 crc kubenswrapper[4888]: I1124 02:13:03.907175 4888 generic.go:334] "Generic (PLEG): container finished" podID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerID="5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a" exitCode=0 Nov 24 02:13:03 crc kubenswrapper[4888]: I1124 02:13:03.907232 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerDied","Data":"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a"} Nov 24 02:13:04 crc kubenswrapper[4888]: I1124 02:13:04.918928 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerStarted","Data":"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944"} Nov 24 02:13:04 crc kubenswrapper[4888]: I1124 02:13:04.961941 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r4w94" podStartSLOduration=2.471335513 podStartE2EDuration="4.961920778s" podCreationTimestamp="2025-11-24 02:13:00 +0000 UTC" firstStartedPulling="2025-11-24 02:13:01.868492941 +0000 UTC m=+6484.451176995" lastFinishedPulling="2025-11-24 02:13:04.359078206 +0000 UTC m=+6486.941762260" observedRunningTime="2025-11-24 02:13:04.96056082 +0000 UTC m=+6487.543244864" watchObservedRunningTime="2025-11-24 02:13:04.961920778 +0000 UTC m=+6487.544604832" Nov 24 02:13:09 crc kubenswrapper[4888]: I1124 02:13:09.245307 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:13:09 crc kubenswrapper[4888]: E1124 02:13:09.245653 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:13:10 crc kubenswrapper[4888]: I1124 02:13:10.663532 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:10 crc kubenswrapper[4888]: I1124 02:13:10.663607 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:10 crc kubenswrapper[4888]: I1124 02:13:10.727104 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:11 crc kubenswrapper[4888]: I1124 02:13:11.050371 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:11 crc kubenswrapper[4888]: I1124 02:13:11.106945 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.022293 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r4w94" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="registry-server" containerID="cri-o://1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944" gracePeriod=2 Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.583449 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.740481 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities\") pod \"8ce11338-c975-40c1-ac2c-3818a542f3a4\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.740554 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content\") pod \"8ce11338-c975-40c1-ac2c-3818a542f3a4\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.740645 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgt2k\" (UniqueName: \"kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k\") pod \"8ce11338-c975-40c1-ac2c-3818a542f3a4\" (UID: \"8ce11338-c975-40c1-ac2c-3818a542f3a4\") " Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.741678 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities" (OuterVolumeSpecName: "utilities") pod "8ce11338-c975-40c1-ac2c-3818a542f3a4" (UID: "8ce11338-c975-40c1-ac2c-3818a542f3a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.747350 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k" (OuterVolumeSpecName: "kube-api-access-zgt2k") pod "8ce11338-c975-40c1-ac2c-3818a542f3a4" (UID: "8ce11338-c975-40c1-ac2c-3818a542f3a4"). InnerVolumeSpecName "kube-api-access-zgt2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.792535 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ce11338-c975-40c1-ac2c-3818a542f3a4" (UID: "8ce11338-c975-40c1-ac2c-3818a542f3a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.843338 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.843398 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce11338-c975-40c1-ac2c-3818a542f3a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:13:13 crc kubenswrapper[4888]: I1124 02:13:13.843409 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgt2k\" (UniqueName: \"kubernetes.io/projected/8ce11338-c975-40c1-ac2c-3818a542f3a4-kube-api-access-zgt2k\") on node \"crc\" DevicePath \"\"" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.033313 4888 generic.go:334] "Generic (PLEG): container finished" podID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerID="1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944" exitCode=0 Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.033357 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerDied","Data":"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944"} Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.033391 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4w94" event={"ID":"8ce11338-c975-40c1-ac2c-3818a542f3a4","Type":"ContainerDied","Data":"f39fad44bb827202f97564c2a258402fa4b0c9cf3a22396b2c23dfcb311e039c"} Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.033409 4888 scope.go:117] "RemoveContainer" containerID="1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.033562 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4w94" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.069378 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.076151 4888 scope.go:117] "RemoveContainer" containerID="5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.080088 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r4w94"] Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.105614 4888 scope.go:117] "RemoveContainer" containerID="6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.165039 4888 scope.go:117] "RemoveContainer" containerID="1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944" Nov 24 02:13:14 crc kubenswrapper[4888]: E1124 02:13:14.165561 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944\": container with ID starting with 1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944 not found: ID does not exist" containerID="1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.165599 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944"} err="failed to get container status \"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944\": rpc error: code = NotFound desc = could not find container \"1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944\": container with ID starting with 1975791e337575922a94ce56d49674fb9faa55b0d225a171b88c371458175944 not found: ID does not exist" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.165618 4888 scope.go:117] "RemoveContainer" containerID="5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a" Nov 24 02:13:14 crc kubenswrapper[4888]: E1124 02:13:14.165948 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a\": container with ID starting with 5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a not found: ID does not exist" containerID="5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.165974 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a"} err="failed to get container status \"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a\": rpc error: code = NotFound desc = could not find container \"5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a\": container with ID starting with 5be73b9c1a00a84644ddb4eb88ff35a9b2bf33537b583a1d4b32021a39959d1a not found: ID does not exist" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.165992 4888 scope.go:117] "RemoveContainer" containerID="6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e" Nov 24 02:13:14 crc kubenswrapper[4888]: E1124 02:13:14.166357 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e\": container with ID starting with 6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e not found: ID does not exist" containerID="6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.166381 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e"} err="failed to get container status \"6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e\": rpc error: code = NotFound desc = could not find container \"6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e\": container with ID starting with 6b8008eeaab6615d33264c632ce16a84b4d8615460a8eacaaac5183b95b5e83e not found: ID does not exist" Nov 24 02:13:14 crc kubenswrapper[4888]: I1124 02:13:14.261736 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" path="/var/lib/kubelet/pods/8ce11338-c975-40c1-ac2c-3818a542f3a4/volumes" Nov 24 02:13:21 crc kubenswrapper[4888]: I1124 02:13:21.245103 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:13:21 crc kubenswrapper[4888]: E1124 02:13:21.246005 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:13:32 crc kubenswrapper[4888]: I1124 02:13:32.245834 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:13:32 crc kubenswrapper[4888]: E1124 02:13:32.246597 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:13:43 crc kubenswrapper[4888]: I1124 02:13:43.245375 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:13:43 crc kubenswrapper[4888]: E1124 02:13:43.246292 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:13:56 crc kubenswrapper[4888]: I1124 02:13:56.246012 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:13:56 crc kubenswrapper[4888]: E1124 02:13:56.248043 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:14:07 crc kubenswrapper[4888]: I1124 02:14:07.245740 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:14:07 crc kubenswrapper[4888]: E1124 02:14:07.247027 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:14:19 crc kubenswrapper[4888]: I1124 02:14:19.245624 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:14:19 crc kubenswrapper[4888]: E1124 02:14:19.246538 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.021253 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:34 crc kubenswrapper[4888]: E1124 02:14:34.022493 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="extract-utilities" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.022511 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="extract-utilities" Nov 24 02:14:34 crc kubenswrapper[4888]: E1124 02:14:34.022533 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="extract-content" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.022541 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="extract-content" Nov 24 02:14:34 crc kubenswrapper[4888]: E1124 02:14:34.022557 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="registry-server" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.022564 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="registry-server" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.022866 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce11338-c975-40c1-ac2c-3818a542f3a4" containerName="registry-server" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.025234 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.055459 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.094779 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnkf\" (UniqueName: \"kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.094969 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.095041 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.198008 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnkf\" (UniqueName: \"kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.198206 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.198258 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.198833 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.198926 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.222707 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnkf\" (UniqueName: \"kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf\") pod \"redhat-marketplace-sgwt7\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.245706 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:14:34 crc kubenswrapper[4888]: E1124 02:14:34.246080 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.358581 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:34 crc kubenswrapper[4888]: I1124 02:14:34.828448 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:35 crc kubenswrapper[4888]: I1124 02:14:35.054379 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerStarted","Data":"9d8493cd311b9ee128df6041d9cc29f05b9eafb8388a5504670fe39b080f50ce"} Nov 24 02:14:36 crc kubenswrapper[4888]: I1124 02:14:36.067330 4888 generic.go:334] "Generic (PLEG): container finished" podID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerID="1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7" exitCode=0 Nov 24 02:14:36 crc kubenswrapper[4888]: I1124 02:14:36.067420 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerDied","Data":"1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7"} Nov 24 02:14:36 crc kubenswrapper[4888]: I1124 02:14:36.071729 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:14:38 crc kubenswrapper[4888]: I1124 02:14:38.093898 4888 generic.go:334] "Generic (PLEG): container finished" podID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerID="ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81" exitCode=0 Nov 24 02:14:38 crc kubenswrapper[4888]: I1124 02:14:38.093985 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerDied","Data":"ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81"} Nov 24 02:14:39 crc kubenswrapper[4888]: I1124 02:14:39.109437 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerStarted","Data":"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e"} Nov 24 02:14:39 crc kubenswrapper[4888]: I1124 02:14:39.133761 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sgwt7" podStartSLOduration=3.704083617 podStartE2EDuration="6.13373884s" podCreationTimestamp="2025-11-24 02:14:33 +0000 UTC" firstStartedPulling="2025-11-24 02:14:36.07153587 +0000 UTC m=+6578.654219914" lastFinishedPulling="2025-11-24 02:14:38.501191093 +0000 UTC m=+6581.083875137" observedRunningTime="2025-11-24 02:14:39.130215291 +0000 UTC m=+6581.712899335" watchObservedRunningTime="2025-11-24 02:14:39.13373884 +0000 UTC m=+6581.716422894" Nov 24 02:14:44 crc kubenswrapper[4888]: I1124 02:14:44.359028 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:44 crc kubenswrapper[4888]: I1124 02:14:44.359410 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:44 crc kubenswrapper[4888]: I1124 02:14:44.458312 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:45 crc kubenswrapper[4888]: I1124 02:14:45.248940 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:45 crc kubenswrapper[4888]: I1124 02:14:45.311438 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.213723 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sgwt7" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="registry-server" containerID="cri-o://2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e" gracePeriod=2 Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.867230 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.976438 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bnkf\" (UniqueName: \"kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf\") pod \"c13ab93f-97a5-4121-8a6c-794721f49afc\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.976712 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities\") pod \"c13ab93f-97a5-4121-8a6c-794721f49afc\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.976891 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content\") pod \"c13ab93f-97a5-4121-8a6c-794721f49afc\" (UID: \"c13ab93f-97a5-4121-8a6c-794721f49afc\") " Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.978474 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities" (OuterVolumeSpecName: "utilities") pod "c13ab93f-97a5-4121-8a6c-794721f49afc" (UID: "c13ab93f-97a5-4121-8a6c-794721f49afc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.984036 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf" (OuterVolumeSpecName: "kube-api-access-7bnkf") pod "c13ab93f-97a5-4121-8a6c-794721f49afc" (UID: "c13ab93f-97a5-4121-8a6c-794721f49afc"). InnerVolumeSpecName "kube-api-access-7bnkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:14:47 crc kubenswrapper[4888]: I1124 02:14:47.997864 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c13ab93f-97a5-4121-8a6c-794721f49afc" (UID: "c13ab93f-97a5-4121-8a6c-794721f49afc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.082018 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.082075 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ab93f-97a5-4121-8a6c-794721f49afc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.082102 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bnkf\" (UniqueName: \"kubernetes.io/projected/c13ab93f-97a5-4121-8a6c-794721f49afc-kube-api-access-7bnkf\") on node \"crc\" DevicePath \"\"" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.232456 4888 generic.go:334] "Generic (PLEG): container finished" podID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerID="2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e" exitCode=0 Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.232517 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerDied","Data":"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e"} Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.232557 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgwt7" event={"ID":"c13ab93f-97a5-4121-8a6c-794721f49afc","Type":"ContainerDied","Data":"9d8493cd311b9ee128df6041d9cc29f05b9eafb8388a5504670fe39b080f50ce"} Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.232582 4888 scope.go:117] "RemoveContainer" containerID="2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.232601 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgwt7" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.297714 4888 scope.go:117] "RemoveContainer" containerID="ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.301583 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.312316 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgwt7"] Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.326369 4888 scope.go:117] "RemoveContainer" containerID="1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.402518 4888 scope.go:117] "RemoveContainer" containerID="2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e" Nov 24 02:14:48 crc kubenswrapper[4888]: E1124 02:14:48.403168 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e\": container with ID starting with 2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e not found: ID does not exist" containerID="2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.403257 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e"} err="failed to get container status \"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e\": rpc error: code = NotFound desc = could not find container \"2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e\": container with ID starting with 2f865faefe9a190fcf378fabce50da457e3070a260b8e7295e199355011d535e not found: ID does not exist" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.403325 4888 scope.go:117] "RemoveContainer" containerID="ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81" Nov 24 02:14:48 crc kubenswrapper[4888]: E1124 02:14:48.403805 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81\": container with ID starting with ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81 not found: ID does not exist" containerID="ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.403939 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81"} err="failed to get container status \"ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81\": rpc error: code = NotFound desc = could not find container \"ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81\": container with ID starting with ff5f5750edbd96c1b4451e73b3ce83e6f5615ef92b7d5291db8245952b111e81 not found: ID does not exist" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.403995 4888 scope.go:117] "RemoveContainer" containerID="1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7" Nov 24 02:14:48 crc kubenswrapper[4888]: E1124 02:14:48.404760 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7\": container with ID starting with 1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7 not found: ID does not exist" containerID="1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7" Nov 24 02:14:48 crc kubenswrapper[4888]: I1124 02:14:48.404845 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7"} err="failed to get container status \"1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7\": rpc error: code = NotFound desc = could not find container \"1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7\": container with ID starting with 1785608c2803b2e24c65c1a273df8c3ef1b8f52bee2ed08b3a216eb64e9b9df7 not found: ID does not exist" Nov 24 02:14:49 crc kubenswrapper[4888]: I1124 02:14:49.251486 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:14:49 crc kubenswrapper[4888]: E1124 02:14:49.252903 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:14:50 crc kubenswrapper[4888]: I1124 02:14:50.260708 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" path="/var/lib/kubelet/pods/c13ab93f-97a5-4121-8a6c-794721f49afc/volumes" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.237795 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb"] Nov 24 02:15:00 crc kubenswrapper[4888]: E1124 02:15:00.238919 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="extract-utilities" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.238938 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="extract-utilities" Nov 24 02:15:00 crc kubenswrapper[4888]: E1124 02:15:00.238956 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="registry-server" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.238963 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="registry-server" Nov 24 02:15:00 crc kubenswrapper[4888]: E1124 02:15:00.238991 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="extract-content" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.239000 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="extract-content" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.239276 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13ab93f-97a5-4121-8a6c-794721f49afc" containerName="registry-server" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.240090 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.244778 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:15:00 crc kubenswrapper[4888]: E1124 02:15:00.245128 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.253887 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.257280 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.267965 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb"] Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.300468 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvf9j\" (UniqueName: \"kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.300543 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.300706 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.402396 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvf9j\" (UniqueName: \"kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.402469 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.402585 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.404303 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.415497 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.420403 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvf9j\" (UniqueName: \"kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j\") pod \"collect-profiles-29399175-qrfcb\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:00 crc kubenswrapper[4888]: I1124 02:15:00.571726 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:01 crc kubenswrapper[4888]: I1124 02:15:01.124440 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb"] Nov 24 02:15:01 crc kubenswrapper[4888]: I1124 02:15:01.417308 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" event={"ID":"5ef5ee64-63d6-4352-9c06-44f3a51d76e6","Type":"ContainerStarted","Data":"576a9b48b1ba4c5871d2f856621e2af3c282e0302b7cfb8f3e1edfd370aa02b6"} Nov 24 02:15:01 crc kubenswrapper[4888]: I1124 02:15:01.417351 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" event={"ID":"5ef5ee64-63d6-4352-9c06-44f3a51d76e6","Type":"ContainerStarted","Data":"e9ed3ca5ffda324ad476fa804ed62f6d8af8db369b9daab8437f934ec26a508d"} Nov 24 02:15:01 crc kubenswrapper[4888]: I1124 02:15:01.434189 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" podStartSLOduration=1.434172029 podStartE2EDuration="1.434172029s" podCreationTimestamp="2025-11-24 02:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:15:01.432157632 +0000 UTC m=+6604.014841686" watchObservedRunningTime="2025-11-24 02:15:01.434172029 +0000 UTC m=+6604.016856073" Nov 24 02:15:03 crc kubenswrapper[4888]: I1124 02:15:03.438086 4888 generic.go:334] "Generic (PLEG): container finished" podID="5ef5ee64-63d6-4352-9c06-44f3a51d76e6" containerID="576a9b48b1ba4c5871d2f856621e2af3c282e0302b7cfb8f3e1edfd370aa02b6" exitCode=0 Nov 24 02:15:03 crc kubenswrapper[4888]: I1124 02:15:03.438268 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" event={"ID":"5ef5ee64-63d6-4352-9c06-44f3a51d76e6","Type":"ContainerDied","Data":"576a9b48b1ba4c5871d2f856621e2af3c282e0302b7cfb8f3e1edfd370aa02b6"} Nov 24 02:15:04 crc kubenswrapper[4888]: I1124 02:15:04.961934 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.011058 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume\") pod \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.011350 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvf9j\" (UniqueName: \"kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j\") pod \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.011408 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume\") pod \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\" (UID: \"5ef5ee64-63d6-4352-9c06-44f3a51d76e6\") " Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.014054 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5ef5ee64-63d6-4352-9c06-44f3a51d76e6" (UID: "5ef5ee64-63d6-4352-9c06-44f3a51d76e6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.017336 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5ef5ee64-63d6-4352-9c06-44f3a51d76e6" (UID: "5ef5ee64-63d6-4352-9c06-44f3a51d76e6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.017370 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j" (OuterVolumeSpecName: "kube-api-access-bvf9j") pod "5ef5ee64-63d6-4352-9c06-44f3a51d76e6" (UID: "5ef5ee64-63d6-4352-9c06-44f3a51d76e6"). InnerVolumeSpecName "kube-api-access-bvf9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.114305 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvf9j\" (UniqueName: \"kubernetes.io/projected/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-kube-api-access-bvf9j\") on node \"crc\" DevicePath \"\"" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.114338 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.114347 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef5ee64-63d6-4352-9c06-44f3a51d76e6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.472168 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" event={"ID":"5ef5ee64-63d6-4352-9c06-44f3a51d76e6","Type":"ContainerDied","Data":"e9ed3ca5ffda324ad476fa804ed62f6d8af8db369b9daab8437f934ec26a508d"} Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.472645 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9ed3ca5ffda324ad476fa804ed62f6d8af8db369b9daab8437f934ec26a508d" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.472271 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399175-qrfcb" Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.580270 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj"] Nov 24 02:15:05 crc kubenswrapper[4888]: I1124 02:15:05.596512 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399130-g9hsj"] Nov 24 02:15:06 crc kubenswrapper[4888]: I1124 02:15:06.287018 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9ece5e-18a0-4880-ad86-1764daf3584a" path="/var/lib/kubelet/pods/3d9ece5e-18a0-4880-ad86-1764daf3584a/volumes" Nov 24 02:15:13 crc kubenswrapper[4888]: I1124 02:15:13.245974 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:15:13 crc kubenswrapper[4888]: E1124 02:15:13.247181 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:15:22 crc kubenswrapper[4888]: I1124 02:15:22.272450 4888 scope.go:117] "RemoveContainer" containerID="da4d8a886c63611947a0e9119a3c94a4c7aee5d3c20e4b8b475fb834ec9da1d0" Nov 24 02:15:27 crc kubenswrapper[4888]: I1124 02:15:27.245056 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:15:27 crc kubenswrapper[4888]: I1124 02:15:27.755114 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415"} Nov 24 02:17:53 crc kubenswrapper[4888]: I1124 02:17:53.591485 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:17:53 crc kubenswrapper[4888]: I1124 02:17:53.592486 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.604682 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:11 crc kubenswrapper[4888]: E1124 02:18:11.606452 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef5ee64-63d6-4352-9c06-44f3a51d76e6" containerName="collect-profiles" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.606475 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef5ee64-63d6-4352-9c06-44f3a51d76e6" containerName="collect-profiles" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.606874 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef5ee64-63d6-4352-9c06-44f3a51d76e6" containerName="collect-profiles" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.615512 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.619061 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.749426 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt8s5\" (UniqueName: \"kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.749527 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.749596 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.851715 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt8s5\" (UniqueName: \"kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.851863 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.851989 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.852925 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.852937 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.878341 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt8s5\" (UniqueName: \"kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5\") pod \"community-operators-mq67j\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:11 crc kubenswrapper[4888]: I1124 02:18:11.956530 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:12 crc kubenswrapper[4888]: I1124 02:18:12.446236 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:12 crc kubenswrapper[4888]: W1124 02:18:12.454259 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00582380_62de_4882_a98e_a09f2926b357.slice/crio-23984174f7beb76d2bdae05aa5e4a5e71048dbc254a5d546edbfd117eef0d7fe WatchSource:0}: Error finding container 23984174f7beb76d2bdae05aa5e4a5e71048dbc254a5d546edbfd117eef0d7fe: Status 404 returned error can't find the container with id 23984174f7beb76d2bdae05aa5e4a5e71048dbc254a5d546edbfd117eef0d7fe Nov 24 02:18:12 crc kubenswrapper[4888]: I1124 02:18:12.800969 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerStarted","Data":"23984174f7beb76d2bdae05aa5e4a5e71048dbc254a5d546edbfd117eef0d7fe"} Nov 24 02:18:13 crc kubenswrapper[4888]: I1124 02:18:13.814890 4888 generic.go:334] "Generic (PLEG): container finished" podID="00582380-62de-4882-a98e-a09f2926b357" containerID="ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9" exitCode=0 Nov 24 02:18:13 crc kubenswrapper[4888]: I1124 02:18:13.815058 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerDied","Data":"ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9"} Nov 24 02:18:15 crc kubenswrapper[4888]: I1124 02:18:15.840398 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerStarted","Data":"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf"} Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.785869 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.789661 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.796853 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.852086 4888 generic.go:334] "Generic (PLEG): container finished" podID="00582380-62de-4882-a98e-a09f2926b357" containerID="7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf" exitCode=0 Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.852144 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerDied","Data":"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf"} Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.989576 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.989962 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:16 crc kubenswrapper[4888]: I1124 02:18:16.990083 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rzvm\" (UniqueName: \"kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.091861 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rzvm\" (UniqueName: \"kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.092029 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.092092 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.093325 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.093407 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.113541 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rzvm\" (UniqueName: \"kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm\") pod \"redhat-operators-dfxxt\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.115993 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.605868 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.880894 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerStarted","Data":"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d"} Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.891553 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerStarted","Data":"6556a9b0a4abb5eebb47e6d301e35c45f4c7abc81364f15157cd8669d07fcaa3"} Nov 24 02:18:17 crc kubenswrapper[4888]: I1124 02:18:17.911058 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mq67j" podStartSLOduration=3.469223698 podStartE2EDuration="6.911036577s" podCreationTimestamp="2025-11-24 02:18:11 +0000 UTC" firstStartedPulling="2025-11-24 02:18:13.817224575 +0000 UTC m=+6796.399908629" lastFinishedPulling="2025-11-24 02:18:17.259037464 +0000 UTC m=+6799.841721508" observedRunningTime="2025-11-24 02:18:17.897742453 +0000 UTC m=+6800.480426497" watchObservedRunningTime="2025-11-24 02:18:17.911036577 +0000 UTC m=+6800.493720621" Nov 24 02:18:18 crc kubenswrapper[4888]: E1124 02:18:18.063380 4888 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d5abf08_b521_47f2_a89f_ce5c9cdcf36d.slice/crio-conmon-7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7.scope\": RecentStats: unable to find data in memory cache]" Nov 24 02:18:18 crc kubenswrapper[4888]: I1124 02:18:18.905409 4888 generic.go:334] "Generic (PLEG): container finished" podID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerID="7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7" exitCode=0 Nov 24 02:18:18 crc kubenswrapper[4888]: I1124 02:18:18.905509 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerDied","Data":"7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7"} Nov 24 02:18:20 crc kubenswrapper[4888]: I1124 02:18:20.931472 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerStarted","Data":"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16"} Nov 24 02:18:21 crc kubenswrapper[4888]: I1124 02:18:21.957470 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:21 crc kubenswrapper[4888]: I1124 02:18:21.957929 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:22 crc kubenswrapper[4888]: I1124 02:18:22.025433 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:23 crc kubenswrapper[4888]: I1124 02:18:23.005913 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:23 crc kubenswrapper[4888]: I1124 02:18:23.591959 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:18:23 crc kubenswrapper[4888]: I1124 02:18:23.592292 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:18:23 crc kubenswrapper[4888]: I1124 02:18:23.969993 4888 generic.go:334] "Generic (PLEG): container finished" podID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerID="27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16" exitCode=0 Nov 24 02:18:23 crc kubenswrapper[4888]: I1124 02:18:23.970061 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerDied","Data":"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16"} Nov 24 02:18:24 crc kubenswrapper[4888]: I1124 02:18:24.979390 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:24 crc kubenswrapper[4888]: I1124 02:18:24.986387 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerStarted","Data":"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8"} Nov 24 02:18:24 crc kubenswrapper[4888]: I1124 02:18:24.986567 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mq67j" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="registry-server" containerID="cri-o://945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d" gracePeriod=2 Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.016460 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dfxxt" podStartSLOduration=3.528464903 podStartE2EDuration="9.016435488s" podCreationTimestamp="2025-11-24 02:18:16 +0000 UTC" firstStartedPulling="2025-11-24 02:18:18.909174968 +0000 UTC m=+6801.491859052" lastFinishedPulling="2025-11-24 02:18:24.397145553 +0000 UTC m=+6806.979829637" observedRunningTime="2025-11-24 02:18:25.005218403 +0000 UTC m=+6807.587902447" watchObservedRunningTime="2025-11-24 02:18:25.016435488 +0000 UTC m=+6807.599119572" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.532028 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.625216 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content\") pod \"00582380-62de-4882-a98e-a09f2926b357\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.626008 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt8s5\" (UniqueName: \"kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5\") pod \"00582380-62de-4882-a98e-a09f2926b357\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.626448 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities\") pod \"00582380-62de-4882-a98e-a09f2926b357\" (UID: \"00582380-62de-4882-a98e-a09f2926b357\") " Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.628119 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities" (OuterVolumeSpecName: "utilities") pod "00582380-62de-4882-a98e-a09f2926b357" (UID: "00582380-62de-4882-a98e-a09f2926b357"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.634511 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5" (OuterVolumeSpecName: "kube-api-access-jt8s5") pod "00582380-62de-4882-a98e-a09f2926b357" (UID: "00582380-62de-4882-a98e-a09f2926b357"). InnerVolumeSpecName "kube-api-access-jt8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.677455 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00582380-62de-4882-a98e-a09f2926b357" (UID: "00582380-62de-4882-a98e-a09f2926b357"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.729398 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.729427 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582380-62de-4882-a98e-a09f2926b357-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:25 crc kubenswrapper[4888]: I1124 02:18:25.729439 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt8s5\" (UniqueName: \"kubernetes.io/projected/00582380-62de-4882-a98e-a09f2926b357-kube-api-access-jt8s5\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.001151 4888 generic.go:334] "Generic (PLEG): container finished" podID="00582380-62de-4882-a98e-a09f2926b357" containerID="945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d" exitCode=0 Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.001192 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerDied","Data":"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d"} Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.001229 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mq67j" event={"ID":"00582380-62de-4882-a98e-a09f2926b357","Type":"ContainerDied","Data":"23984174f7beb76d2bdae05aa5e4a5e71048dbc254a5d546edbfd117eef0d7fe"} Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.001246 4888 scope.go:117] "RemoveContainer" containerID="945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.002095 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mq67j" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.026002 4888 scope.go:117] "RemoveContainer" containerID="7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.047101 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.057581 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mq67j"] Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.073023 4888 scope.go:117] "RemoveContainer" containerID="ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.129828 4888 scope.go:117] "RemoveContainer" containerID="945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d" Nov 24 02:18:26 crc kubenswrapper[4888]: E1124 02:18:26.130600 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d\": container with ID starting with 945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d not found: ID does not exist" containerID="945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.130629 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d"} err="failed to get container status \"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d\": rpc error: code = NotFound desc = could not find container \"945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d\": container with ID starting with 945e0896d20165e27e5be4696a0b636b4dc6808ba461aff447194bdf837f684d not found: ID does not exist" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.130652 4888 scope.go:117] "RemoveContainer" containerID="7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf" Nov 24 02:18:26 crc kubenswrapper[4888]: E1124 02:18:26.131076 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf\": container with ID starting with 7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf not found: ID does not exist" containerID="7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.131106 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf"} err="failed to get container status \"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf\": rpc error: code = NotFound desc = could not find container \"7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf\": container with ID starting with 7e6147ecb84a394b231a41999fbbc030c4de1e6700aab72ec04877041b3067cf not found: ID does not exist" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.131122 4888 scope.go:117] "RemoveContainer" containerID="ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9" Nov 24 02:18:26 crc kubenswrapper[4888]: E1124 02:18:26.131423 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9\": container with ID starting with ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9 not found: ID does not exist" containerID="ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.131470 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9"} err="failed to get container status \"ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9\": rpc error: code = NotFound desc = could not find container \"ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9\": container with ID starting with ed0424d7fbd7e4962fd17afd2952425f275c1ec622f39998f75c5185eb6b53f9 not found: ID does not exist" Nov 24 02:18:26 crc kubenswrapper[4888]: I1124 02:18:26.266393 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00582380-62de-4882-a98e-a09f2926b357" path="/var/lib/kubelet/pods/00582380-62de-4882-a98e-a09f2926b357/volumes" Nov 24 02:18:27 crc kubenswrapper[4888]: I1124 02:18:27.116547 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:27 crc kubenswrapper[4888]: I1124 02:18:27.116640 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:28 crc kubenswrapper[4888]: I1124 02:18:28.190902 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dfxxt" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" probeResult="failure" output=< Nov 24 02:18:28 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:18:28 crc kubenswrapper[4888]: > Nov 24 02:18:38 crc kubenswrapper[4888]: I1124 02:18:38.170986 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dfxxt" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" probeResult="failure" output=< Nov 24 02:18:38 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:18:38 crc kubenswrapper[4888]: > Nov 24 02:18:47 crc kubenswrapper[4888]: I1124 02:18:47.180446 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:47 crc kubenswrapper[4888]: I1124 02:18:47.279755 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:48 crc kubenswrapper[4888]: I1124 02:18:48.082156 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:48 crc kubenswrapper[4888]: I1124 02:18:48.294054 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dfxxt" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" containerID="cri-o://c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8" gracePeriod=2 Nov 24 02:18:48 crc kubenswrapper[4888]: I1124 02:18:48.875938 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.023135 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities\") pod \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.023251 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rzvm\" (UniqueName: \"kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm\") pod \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.023290 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") pod \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.024651 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities" (OuterVolumeSpecName: "utilities") pod "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" (UID: "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.032613 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm" (OuterVolumeSpecName: "kube-api-access-9rzvm") pod "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" (UID: "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d"). InnerVolumeSpecName "kube-api-access-9rzvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.125121 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" (UID: "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.127298 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") pod \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\" (UID: \"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d\") " Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.128144 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.128248 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rzvm\" (UniqueName: \"kubernetes.io/projected/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-kube-api-access-9rzvm\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:49 crc kubenswrapper[4888]: W1124 02:18:49.130061 4888 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d/volumes/kubernetes.io~empty-dir/catalog-content Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.131595 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" (UID: "4d5abf08-b521-47f2-a89f-ce5c9cdcf36d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.231268 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.310050 4888 generic.go:334] "Generic (PLEG): container finished" podID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerID="c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8" exitCode=0 Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.310097 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerDied","Data":"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8"} Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.310125 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfxxt" event={"ID":"4d5abf08-b521-47f2-a89f-ce5c9cdcf36d","Type":"ContainerDied","Data":"6556a9b0a4abb5eebb47e6d301e35c45f4c7abc81364f15157cd8669d07fcaa3"} Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.310145 4888 scope.go:117] "RemoveContainer" containerID="c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.310284 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfxxt" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.348798 4888 scope.go:117] "RemoveContainer" containerID="27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.354636 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.368050 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dfxxt"] Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.375524 4888 scope.go:117] "RemoveContainer" containerID="7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.451382 4888 scope.go:117] "RemoveContainer" containerID="c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8" Nov 24 02:18:49 crc kubenswrapper[4888]: E1124 02:18:49.451903 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8\": container with ID starting with c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8 not found: ID does not exist" containerID="c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.452028 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8"} err="failed to get container status \"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8\": rpc error: code = NotFound desc = could not find container \"c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8\": container with ID starting with c8d4116e63821746cea2028cbf46a4daa4d511a923dbb617df7ce39bbdc600f8 not found: ID does not exist" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.452135 4888 scope.go:117] "RemoveContainer" containerID="27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16" Nov 24 02:18:49 crc kubenswrapper[4888]: E1124 02:18:49.452778 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16\": container with ID starting with 27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16 not found: ID does not exist" containerID="27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.452808 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16"} err="failed to get container status \"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16\": rpc error: code = NotFound desc = could not find container \"27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16\": container with ID starting with 27aea567186f346a26816efe8a8a6e5b3449b04856cd4b268553892ad9b6ab16 not found: ID does not exist" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.452847 4888 scope.go:117] "RemoveContainer" containerID="7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7" Nov 24 02:18:49 crc kubenswrapper[4888]: E1124 02:18:49.453102 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7\": container with ID starting with 7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7 not found: ID does not exist" containerID="7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7" Nov 24 02:18:49 crc kubenswrapper[4888]: I1124 02:18:49.453123 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7"} err="failed to get container status \"7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7\": rpc error: code = NotFound desc = could not find container \"7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7\": container with ID starting with 7a485802d6bf81c7e483af88ffd126e0024956e8bf206104061c16856467e7b7 not found: ID does not exist" Nov 24 02:18:50 crc kubenswrapper[4888]: I1124 02:18:50.266260 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" path="/var/lib/kubelet/pods/4d5abf08-b521-47f2-a89f-ce5c9cdcf36d/volumes" Nov 24 02:18:53 crc kubenswrapper[4888]: I1124 02:18:53.591475 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:18:53 crc kubenswrapper[4888]: I1124 02:18:53.592514 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:18:53 crc kubenswrapper[4888]: I1124 02:18:53.592591 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:18:53 crc kubenswrapper[4888]: I1124 02:18:53.594115 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:18:53 crc kubenswrapper[4888]: I1124 02:18:53.594221 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415" gracePeriod=600 Nov 24 02:18:54 crc kubenswrapper[4888]: I1124 02:18:54.385676 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415" exitCode=0 Nov 24 02:18:54 crc kubenswrapper[4888]: I1124 02:18:54.385744 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415"} Nov 24 02:18:54 crc kubenswrapper[4888]: I1124 02:18:54.386318 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6"} Nov 24 02:18:54 crc kubenswrapper[4888]: I1124 02:18:54.386361 4888 scope.go:117] "RemoveContainer" containerID="4e7b3522d47c3750be1f81105a93a268d8106433472a2aa92211f936ba53fa0f" Nov 24 02:20:53 crc kubenswrapper[4888]: I1124 02:20:53.592042 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:20:53 crc kubenswrapper[4888]: I1124 02:20:53.592776 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:21:23 crc kubenswrapper[4888]: I1124 02:21:23.591324 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:21:23 crc kubenswrapper[4888]: I1124 02:21:23.591872 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.591667 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.592713 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.592854 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.594253 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.594365 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" gracePeriod=600 Nov 24 02:21:53 crc kubenswrapper[4888]: E1124 02:21:53.725350 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.843424 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" exitCode=0 Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.843533 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6"} Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.843657 4888 scope.go:117] "RemoveContainer" containerID="6cd2f76103af86942f1ebc38f33f96e4f20dfc14276dec09beb9327f3f919415" Nov 24 02:21:53 crc kubenswrapper[4888]: I1124 02:21:53.844624 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:21:53 crc kubenswrapper[4888]: E1124 02:21:53.847550 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:22:06 crc kubenswrapper[4888]: I1124 02:22:06.246148 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:22:06 crc kubenswrapper[4888]: E1124 02:22:06.247114 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:22:21 crc kubenswrapper[4888]: I1124 02:22:21.245570 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:22:21 crc kubenswrapper[4888]: E1124 02:22:21.246393 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:22:35 crc kubenswrapper[4888]: I1124 02:22:35.245398 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:22:35 crc kubenswrapper[4888]: E1124 02:22:35.246061 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:22:46 crc kubenswrapper[4888]: I1124 02:22:46.245266 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:22:46 crc kubenswrapper[4888]: E1124 02:22:46.246493 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:00 crc kubenswrapper[4888]: I1124 02:23:00.245349 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:23:00 crc kubenswrapper[4888]: E1124 02:23:00.246056 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:14 crc kubenswrapper[4888]: I1124 02:23:14.246657 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:23:14 crc kubenswrapper[4888]: E1124 02:23:14.247851 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:29 crc kubenswrapper[4888]: I1124 02:23:29.246236 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:23:29 crc kubenswrapper[4888]: E1124 02:23:29.247415 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.067254 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068458 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068480 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068519 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="extract-content" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068533 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="extract-content" Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068564 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068577 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068603 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="extract-utilities" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068615 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="extract-utilities" Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068700 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="extract-utilities" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068712 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="extract-utilities" Nov 24 02:23:40 crc kubenswrapper[4888]: E1124 02:23:40.068733 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="extract-content" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.068745 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="extract-content" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.069213 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d5abf08-b521-47f2-a89f-ce5c9cdcf36d" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.069240 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="00582380-62de-4882-a98e-a09f2926b357" containerName="registry-server" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.075000 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.087039 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.194834 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.194895 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.194921 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4lgn\" (UniqueName: \"kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.297072 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.297148 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.297177 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4lgn\" (UniqueName: \"kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.297670 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.297687 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.321791 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4lgn\" (UniqueName: \"kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn\") pod \"certified-operators-psrk9\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.414187 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:40 crc kubenswrapper[4888]: I1124 02:23:40.920050 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:41 crc kubenswrapper[4888]: I1124 02:23:41.223080 4888 generic.go:334] "Generic (PLEG): container finished" podID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerID="62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410" exitCode=0 Nov 24 02:23:41 crc kubenswrapper[4888]: I1124 02:23:41.223146 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerDied","Data":"62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410"} Nov 24 02:23:41 crc kubenswrapper[4888]: I1124 02:23:41.223185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerStarted","Data":"88c5320fc740bebe09d0d7ee958007c2e6bc673b43b5b3567fc32c4f64cd9b65"} Nov 24 02:23:41 crc kubenswrapper[4888]: I1124 02:23:41.225420 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:23:42 crc kubenswrapper[4888]: I1124 02:23:42.241540 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerStarted","Data":"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a"} Nov 24 02:23:43 crc kubenswrapper[4888]: I1124 02:23:43.245460 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:23:43 crc kubenswrapper[4888]: E1124 02:23:43.246405 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:43 crc kubenswrapper[4888]: I1124 02:23:43.257499 4888 generic.go:334] "Generic (PLEG): container finished" podID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerID="9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a" exitCode=0 Nov 24 02:23:43 crc kubenswrapper[4888]: I1124 02:23:43.257551 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerDied","Data":"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a"} Nov 24 02:23:44 crc kubenswrapper[4888]: I1124 02:23:44.272554 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerStarted","Data":"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e"} Nov 24 02:23:44 crc kubenswrapper[4888]: I1124 02:23:44.293447 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-psrk9" podStartSLOduration=1.848934166 podStartE2EDuration="4.29342636s" podCreationTimestamp="2025-11-24 02:23:40 +0000 UTC" firstStartedPulling="2025-11-24 02:23:41.225122983 +0000 UTC m=+7123.807807027" lastFinishedPulling="2025-11-24 02:23:43.669615167 +0000 UTC m=+7126.252299221" observedRunningTime="2025-11-24 02:23:44.293125222 +0000 UTC m=+7126.875809276" watchObservedRunningTime="2025-11-24 02:23:44.29342636 +0000 UTC m=+7126.876110414" Nov 24 02:23:50 crc kubenswrapper[4888]: I1124 02:23:50.414959 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:50 crc kubenswrapper[4888]: I1124 02:23:50.415807 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:50 crc kubenswrapper[4888]: I1124 02:23:50.495500 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:51 crc kubenswrapper[4888]: I1124 02:23:51.478235 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:51 crc kubenswrapper[4888]: I1124 02:23:51.551322 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:53 crc kubenswrapper[4888]: I1124 02:23:53.412495 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-psrk9" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="registry-server" containerID="cri-o://7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e" gracePeriod=2 Nov 24 02:23:53 crc kubenswrapper[4888]: I1124 02:23:53.962705 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.069419 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities\") pod \"8d6c82d9-c470-46bd-a84b-31c32878382d\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.070031 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content\") pod \"8d6c82d9-c470-46bd-a84b-31c32878382d\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.070237 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4lgn\" (UniqueName: \"kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn\") pod \"8d6c82d9-c470-46bd-a84b-31c32878382d\" (UID: \"8d6c82d9-c470-46bd-a84b-31c32878382d\") " Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.070785 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities" (OuterVolumeSpecName: "utilities") pod "8d6c82d9-c470-46bd-a84b-31c32878382d" (UID: "8d6c82d9-c470-46bd-a84b-31c32878382d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.071518 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.076014 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn" (OuterVolumeSpecName: "kube-api-access-z4lgn") pod "8d6c82d9-c470-46bd-a84b-31c32878382d" (UID: "8d6c82d9-c470-46bd-a84b-31c32878382d"). InnerVolumeSpecName "kube-api-access-z4lgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.114264 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d6c82d9-c470-46bd-a84b-31c32878382d" (UID: "8d6c82d9-c470-46bd-a84b-31c32878382d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.173549 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c82d9-c470-46bd-a84b-31c32878382d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.173583 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4lgn\" (UniqueName: \"kubernetes.io/projected/8d6c82d9-c470-46bd-a84b-31c32878382d-kube-api-access-z4lgn\") on node \"crc\" DevicePath \"\"" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.246376 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:23:54 crc kubenswrapper[4888]: E1124 02:23:54.246874 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.433332 4888 generic.go:334] "Generic (PLEG): container finished" podID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerID="7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e" exitCode=0 Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.433399 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerDied","Data":"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e"} Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.433429 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psrk9" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.433456 4888 scope.go:117] "RemoveContainer" containerID="7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.433441 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psrk9" event={"ID":"8d6c82d9-c470-46bd-a84b-31c32878382d","Type":"ContainerDied","Data":"88c5320fc740bebe09d0d7ee958007c2e6bc673b43b5b3567fc32c4f64cd9b65"} Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.478401 4888 scope.go:117] "RemoveContainer" containerID="9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.480500 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.497133 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-psrk9"] Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.506143 4888 scope.go:117] "RemoveContainer" containerID="62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.575725 4888 scope.go:117] "RemoveContainer" containerID="7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e" Nov 24 02:23:54 crc kubenswrapper[4888]: E1124 02:23:54.576161 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e\": container with ID starting with 7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e not found: ID does not exist" containerID="7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.576216 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e"} err="failed to get container status \"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e\": rpc error: code = NotFound desc = could not find container \"7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e\": container with ID starting with 7d913ee4eb7cdcc5c8a8b1a1480cc015f6db1683a340a13ada0664a404853f0e not found: ID does not exist" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.576245 4888 scope.go:117] "RemoveContainer" containerID="9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a" Nov 24 02:23:54 crc kubenswrapper[4888]: E1124 02:23:54.576747 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a\": container with ID starting with 9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a not found: ID does not exist" containerID="9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.576788 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a"} err="failed to get container status \"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a\": rpc error: code = NotFound desc = could not find container \"9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a\": container with ID starting with 9b4ab890c56b803f8ad05882565617679e4f48d14426c7723589e02227b4877a not found: ID does not exist" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.576857 4888 scope.go:117] "RemoveContainer" containerID="62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410" Nov 24 02:23:54 crc kubenswrapper[4888]: E1124 02:23:54.577109 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410\": container with ID starting with 62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410 not found: ID does not exist" containerID="62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410" Nov 24 02:23:54 crc kubenswrapper[4888]: I1124 02:23:54.577139 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410"} err="failed to get container status \"62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410\": rpc error: code = NotFound desc = could not find container \"62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410\": container with ID starting with 62bc8057631afa7b42fe1dbb1874f4b418b7f4a0114c9b1304a8b9d57e205410 not found: ID does not exist" Nov 24 02:23:56 crc kubenswrapper[4888]: I1124 02:23:56.269797 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" path="/var/lib/kubelet/pods/8d6c82d9-c470-46bd-a84b-31c32878382d/volumes" Nov 24 02:24:09 crc kubenswrapper[4888]: I1124 02:24:09.245184 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:24:09 crc kubenswrapper[4888]: E1124 02:24:09.246074 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:24:20 crc kubenswrapper[4888]: I1124 02:24:20.246024 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:24:20 crc kubenswrapper[4888]: E1124 02:24:20.247046 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:24:31 crc kubenswrapper[4888]: I1124 02:24:31.245612 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:24:31 crc kubenswrapper[4888]: E1124 02:24:31.250023 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:24:45 crc kubenswrapper[4888]: I1124 02:24:45.246019 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:24:45 crc kubenswrapper[4888]: E1124 02:24:45.247315 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:24:59 crc kubenswrapper[4888]: I1124 02:24:59.245104 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:24:59 crc kubenswrapper[4888]: E1124 02:24:59.246047 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:25:14 crc kubenswrapper[4888]: I1124 02:25:14.245765 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:25:14 crc kubenswrapper[4888]: E1124 02:25:14.246668 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:25:25 crc kubenswrapper[4888]: I1124 02:25:25.245945 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:25:25 crc kubenswrapper[4888]: E1124 02:25:25.247169 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:25:40 crc kubenswrapper[4888]: I1124 02:25:40.245969 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:25:40 crc kubenswrapper[4888]: E1124 02:25:40.246588 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:25:53 crc kubenswrapper[4888]: I1124 02:25:53.245039 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:25:53 crc kubenswrapper[4888]: E1124 02:25:53.245920 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.159504 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:25:57 crc kubenswrapper[4888]: E1124 02:25:57.160673 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="registry-server" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.160692 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="registry-server" Nov 24 02:25:57 crc kubenswrapper[4888]: E1124 02:25:57.160714 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="extract-utilities" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.160722 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="extract-utilities" Nov 24 02:25:57 crc kubenswrapper[4888]: E1124 02:25:57.160745 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="extract-content" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.160753 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="extract-content" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.161080 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6c82d9-c470-46bd-a84b-31c32878382d" containerName="registry-server" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.163307 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.183087 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.268241 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djnrr\" (UniqueName: \"kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.268307 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.268442 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.371018 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djnrr\" (UniqueName: \"kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.371390 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.371972 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.372629 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.373099 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.392896 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djnrr\" (UniqueName: \"kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr\") pod \"redhat-marketplace-t5k4d\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.485203 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.548650 4888 generic.go:334] "Generic (PLEG): container finished" podID="cc931795-8f17-4afd-a16c-264855b1755c" containerID="10ceea90344fa48a36b4bc996c2a9c6e0ad225c08f50143e81749208083fb023" exitCode=0 Nov 24 02:25:57 crc kubenswrapper[4888]: I1124 02:25:57.548700 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cc931795-8f17-4afd-a16c-264855b1755c","Type":"ContainerDied","Data":"10ceea90344fa48a36b4bc996c2a9c6e0ad225c08f50143e81749208083fb023"} Nov 24 02:25:58 crc kubenswrapper[4888]: I1124 02:25:58.000489 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:25:58 crc kubenswrapper[4888]: I1124 02:25:58.558185 4888 generic.go:334] "Generic (PLEG): container finished" podID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerID="9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec" exitCode=0 Nov 24 02:25:58 crc kubenswrapper[4888]: I1124 02:25:58.558252 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerDied","Data":"9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec"} Nov 24 02:25:58 crc kubenswrapper[4888]: I1124 02:25:58.558572 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerStarted","Data":"723fa7306cbfa2c0e37f44d1edd732184dbc8e971aad3e8f59f0325ec6e5af74"} Nov 24 02:25:58 crc kubenswrapper[4888]: I1124 02:25:58.985613 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.115754 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116166 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116202 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116272 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-956zg\" (UniqueName: \"kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116309 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116386 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116447 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116547 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.116584 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data\") pod \"cc931795-8f17-4afd-a16c-264855b1755c\" (UID: \"cc931795-8f17-4afd-a16c-264855b1755c\") " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.117787 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.118244 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data" (OuterVolumeSpecName: "config-data") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.123104 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.123187 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.125373 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg" (OuterVolumeSpecName: "kube-api-access-956zg") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "kube-api-access-956zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.154133 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.155609 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.157365 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.183556 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cc931795-8f17-4afd-a16c-264855b1755c" (UID: "cc931795-8f17-4afd-a16c-264855b1755c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221192 4888 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221247 4888 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221269 4888 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221291 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-956zg\" (UniqueName: \"kubernetes.io/projected/cc931795-8f17-4afd-a16c-264855b1755c-kube-api-access-956zg\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221320 4888 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221341 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221363 4888 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc931795-8f17-4afd-a16c-264855b1755c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221385 4888 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cc931795-8f17-4afd-a16c-264855b1755c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.221406 4888 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc931795-8f17-4afd-a16c-264855b1755c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.259582 4888 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.323051 4888 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.568528 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerStarted","Data":"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb"} Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.570206 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cc931795-8f17-4afd-a16c-264855b1755c","Type":"ContainerDied","Data":"af22224b967ee37b219f01467a10ceaccbbad177e240d97c7d0bb474ae4d8e2e"} Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.570393 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af22224b967ee37b219f01467a10ceaccbbad177e240d97c7d0bb474ae4d8e2e" Nov 24 02:25:59 crc kubenswrapper[4888]: I1124 02:25:59.570240 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 02:26:00 crc kubenswrapper[4888]: I1124 02:26:00.585248 4888 generic.go:334] "Generic (PLEG): container finished" podID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerID="728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb" exitCode=0 Nov 24 02:26:00 crc kubenswrapper[4888]: I1124 02:26:00.585289 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerDied","Data":"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb"} Nov 24 02:26:01 crc kubenswrapper[4888]: I1124 02:26:01.600155 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerStarted","Data":"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755"} Nov 24 02:26:01 crc kubenswrapper[4888]: I1124 02:26:01.633457 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t5k4d" podStartSLOduration=2.190457372 podStartE2EDuration="4.633436223s" podCreationTimestamp="2025-11-24 02:25:57 +0000 UTC" firstStartedPulling="2025-11-24 02:25:58.56027871 +0000 UTC m=+7261.142962754" lastFinishedPulling="2025-11-24 02:26:01.003257561 +0000 UTC m=+7263.585941605" observedRunningTime="2025-11-24 02:26:01.621000223 +0000 UTC m=+7264.203684287" watchObservedRunningTime="2025-11-24 02:26:01.633436223 +0000 UTC m=+7264.216120277" Nov 24 02:26:07 crc kubenswrapper[4888]: I1124 02:26:07.485451 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:07 crc kubenswrapper[4888]: I1124 02:26:07.485893 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:07 crc kubenswrapper[4888]: I1124 02:26:07.572050 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:07 crc kubenswrapper[4888]: I1124 02:26:07.753289 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:07 crc kubenswrapper[4888]: I1124 02:26:07.821680 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:26:08 crc kubenswrapper[4888]: I1124 02:26:08.260703 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:26:08 crc kubenswrapper[4888]: E1124 02:26:08.261581 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:26:09 crc kubenswrapper[4888]: I1124 02:26:09.700110 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t5k4d" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="registry-server" containerID="cri-o://3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755" gracePeriod=2 Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.249195 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.270519 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.271100 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc931795-8f17-4afd-a16c-264855b1755c" containerName="tempest-tests-tempest-tests-runner" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271121 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc931795-8f17-4afd-a16c-264855b1755c" containerName="tempest-tests-tempest-tests-runner" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.271146 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="registry-server" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271426 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="registry-server" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.271454 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="extract-content" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271462 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="extract-content" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.271486 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="extract-utilities" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271492 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="extract-utilities" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271722 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc931795-8f17-4afd-a16c-264855b1755c" containerName="tempest-tests-tempest-tests-runner" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.271750 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerName="registry-server" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.272702 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.278237 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gc5cj" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.283656 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.295719 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content\") pod \"0f771c28-8cc4-457e-9855-70cb8af7dd61\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.296052 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djnrr\" (UniqueName: \"kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr\") pod \"0f771c28-8cc4-457e-9855-70cb8af7dd61\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.296210 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities\") pod \"0f771c28-8cc4-457e-9855-70cb8af7dd61\" (UID: \"0f771c28-8cc4-457e-9855-70cb8af7dd61\") " Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.298287 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities" (OuterVolumeSpecName: "utilities") pod "0f771c28-8cc4-457e-9855-70cb8af7dd61" (UID: "0f771c28-8cc4-457e-9855-70cb8af7dd61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.307105 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr" (OuterVolumeSpecName: "kube-api-access-djnrr") pod "0f771c28-8cc4-457e-9855-70cb8af7dd61" (UID: "0f771c28-8cc4-457e-9855-70cb8af7dd61"). InnerVolumeSpecName "kube-api-access-djnrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.329751 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f771c28-8cc4-457e-9855-70cb8af7dd61" (UID: "0f771c28-8cc4-457e-9855-70cb8af7dd61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.399989 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.400372 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzltp\" (UniqueName: \"kubernetes.io/projected/f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717-kube-api-access-fzltp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.400592 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djnrr\" (UniqueName: \"kubernetes.io/projected/0f771c28-8cc4-457e-9855-70cb8af7dd61-kube-api-access-djnrr\") on node \"crc\" DevicePath \"\"" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.400734 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.400863 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f771c28-8cc4-457e-9855-70cb8af7dd61-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.502939 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.503379 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzltp\" (UniqueName: \"kubernetes.io/projected/f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717-kube-api-access-fzltp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.505173 4888 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.523070 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzltp\" (UniqueName: \"kubernetes.io/projected/f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717-kube-api-access-fzltp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.541627 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.607808 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.734383 4888 generic.go:334] "Generic (PLEG): container finished" podID="0f771c28-8cc4-457e-9855-70cb8af7dd61" containerID="3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755" exitCode=0 Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.734602 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerDied","Data":"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755"} Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.735185 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5k4d" event={"ID":"0f771c28-8cc4-457e-9855-70cb8af7dd61","Type":"ContainerDied","Data":"723fa7306cbfa2c0e37f44d1edd732184dbc8e971aad3e8f59f0325ec6e5af74"} Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.735208 4888 scope.go:117] "RemoveContainer" containerID="3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.734678 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5k4d" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.775336 4888 scope.go:117] "RemoveContainer" containerID="728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.797224 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.808252 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5k4d"] Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.815763 4888 scope.go:117] "RemoveContainer" containerID="9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.895984 4888 scope.go:117] "RemoveContainer" containerID="3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.896585 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755\": container with ID starting with 3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755 not found: ID does not exist" containerID="3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.896650 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755"} err="failed to get container status \"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755\": rpc error: code = NotFound desc = could not find container \"3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755\": container with ID starting with 3c6845fd680eca59ce0ad43c188639714bab8fc322ad57c1f2588d7313916755 not found: ID does not exist" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.896888 4888 scope.go:117] "RemoveContainer" containerID="728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.897346 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb\": container with ID starting with 728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb not found: ID does not exist" containerID="728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.897392 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb"} err="failed to get container status \"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb\": rpc error: code = NotFound desc = could not find container \"728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb\": container with ID starting with 728f7d2ca5df3b80ee872bb72cde7371631d512ddaba72dd36b5daa4b0906acb not found: ID does not exist" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.897423 4888 scope.go:117] "RemoveContainer" containerID="9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec" Nov 24 02:26:10 crc kubenswrapper[4888]: E1124 02:26:10.897758 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec\": container with ID starting with 9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec not found: ID does not exist" containerID="9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec" Nov 24 02:26:10 crc kubenswrapper[4888]: I1124 02:26:10.897787 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec"} err="failed to get container status \"9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec\": rpc error: code = NotFound desc = could not find container \"9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec\": container with ID starting with 9a8bf866c6a3f0993fce59a4fc09a4a4b565a4c6a00e41f6299f9ee679de7cec not found: ID does not exist" Nov 24 02:26:11 crc kubenswrapper[4888]: I1124 02:26:11.121361 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 02:26:11 crc kubenswrapper[4888]: I1124 02:26:11.759128 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717","Type":"ContainerStarted","Data":"9bf14e1293164b849335b38cc4869e3342029e02ababf834dbbd943cff9077c9"} Nov 24 02:26:12 crc kubenswrapper[4888]: I1124 02:26:12.264601 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f771c28-8cc4-457e-9855-70cb8af7dd61" path="/var/lib/kubelet/pods/0f771c28-8cc4-457e-9855-70cb8af7dd61/volumes" Nov 24 02:26:12 crc kubenswrapper[4888]: I1124 02:26:12.773520 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717","Type":"ContainerStarted","Data":"727d28870e2fd023a63792710d5798a9f876b490ab4de531b807a734c0822976"} Nov 24 02:26:12 crc kubenswrapper[4888]: I1124 02:26:12.812723 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.679445598 podStartE2EDuration="2.81269101s" podCreationTimestamp="2025-11-24 02:26:10 +0000 UTC" firstStartedPulling="2025-11-24 02:26:11.125425076 +0000 UTC m=+7273.708109120" lastFinishedPulling="2025-11-24 02:26:12.258670488 +0000 UTC m=+7274.841354532" observedRunningTime="2025-11-24 02:26:12.790422974 +0000 UTC m=+7275.373107058" watchObservedRunningTime="2025-11-24 02:26:12.81269101 +0000 UTC m=+7275.395375114" Nov 24 02:26:19 crc kubenswrapper[4888]: I1124 02:26:19.246422 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:26:19 crc kubenswrapper[4888]: E1124 02:26:19.247529 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:26:33 crc kubenswrapper[4888]: I1124 02:26:33.245144 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:26:33 crc kubenswrapper[4888]: E1124 02:26:33.245869 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.205665 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-467dh/must-gather-jk6g7"] Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.208327 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.214658 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-467dh"/"openshift-service-ca.crt" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.215033 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-467dh"/"default-dockercfg-f4kxf" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.215247 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-467dh"/"kube-root-ca.crt" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.275514 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-467dh/must-gather-jk6g7"] Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.288758 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.289310 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kmqc\" (UniqueName: \"kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.391422 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kmqc\" (UniqueName: \"kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.391615 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.392066 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.417098 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kmqc\" (UniqueName: \"kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc\") pod \"must-gather-jk6g7\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:42 crc kubenswrapper[4888]: I1124 02:26:42.530576 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:26:43 crc kubenswrapper[4888]: I1124 02:26:43.096665 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-467dh/must-gather-jk6g7"] Nov 24 02:26:43 crc kubenswrapper[4888]: I1124 02:26:43.185427 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/must-gather-jk6g7" event={"ID":"ad08b613-2cfe-4f6f-8793-18c3be4885c0","Type":"ContainerStarted","Data":"a6b3914c2a9c0296c47c36351add332af2b90506b42938d2fcdfac5dd649d806"} Nov 24 02:26:48 crc kubenswrapper[4888]: I1124 02:26:48.254405 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:26:48 crc kubenswrapper[4888]: E1124 02:26:48.255095 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:26:51 crc kubenswrapper[4888]: I1124 02:26:51.281263 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/must-gather-jk6g7" event={"ID":"ad08b613-2cfe-4f6f-8793-18c3be4885c0","Type":"ContainerStarted","Data":"aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48"} Nov 24 02:26:51 crc kubenswrapper[4888]: I1124 02:26:51.281714 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/must-gather-jk6g7" event={"ID":"ad08b613-2cfe-4f6f-8793-18c3be4885c0","Type":"ContainerStarted","Data":"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0"} Nov 24 02:26:51 crc kubenswrapper[4888]: I1124 02:26:51.298668 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-467dh/must-gather-jk6g7" podStartSLOduration=1.734727025 podStartE2EDuration="9.29864954s" podCreationTimestamp="2025-11-24 02:26:42 +0000 UTC" firstStartedPulling="2025-11-24 02:26:43.102074155 +0000 UTC m=+7305.684758199" lastFinishedPulling="2025-11-24 02:26:50.66599666 +0000 UTC m=+7313.248680714" observedRunningTime="2025-11-24 02:26:51.29616076 +0000 UTC m=+7313.878844804" watchObservedRunningTime="2025-11-24 02:26:51.29864954 +0000 UTC m=+7313.881333594" Nov 24 02:26:55 crc kubenswrapper[4888]: E1124 02:26:55.241941 4888 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.23:59870->38.102.83.23:41433: write tcp 38.102.83.23:59870->38.102.83.23:41433: write: broken pipe Nov 24 02:26:55 crc kubenswrapper[4888]: E1124 02:26:55.241947 4888 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.23:59870->38.102.83.23:41433: read tcp 38.102.83.23:59870->38.102.83.23:41433: read: connection reset by peer Nov 24 02:26:56 crc kubenswrapper[4888]: I1124 02:26:56.839150 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-467dh/crc-debug-mkvxv"] Nov 24 02:26:56 crc kubenswrapper[4888]: I1124 02:26:56.841548 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:56 crc kubenswrapper[4888]: I1124 02:26:56.985212 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:56 crc kubenswrapper[4888]: I1124 02:26:56.985731 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.087941 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.088101 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.088482 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.109274 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r\") pod \"crc-debug-mkvxv\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.163793 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:26:57 crc kubenswrapper[4888]: I1124 02:26:57.347999 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-mkvxv" event={"ID":"bfebaba2-f323-4eb1-8bbc-93f71344b2ff","Type":"ContainerStarted","Data":"6b86a7e425685a35e0e0f0a1119dd4e2473145016a236afa95d06a9827c0ef23"} Nov 24 02:26:59 crc kubenswrapper[4888]: I1124 02:26:59.245528 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:27:00 crc kubenswrapper[4888]: I1124 02:27:00.384164 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58"} Nov 24 02:27:09 crc kubenswrapper[4888]: I1124 02:27:09.506470 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-mkvxv" event={"ID":"bfebaba2-f323-4eb1-8bbc-93f71344b2ff","Type":"ContainerStarted","Data":"964a4ec403d0d3242e88140569203f1b5eae7cbb458a12fa63e802a3281a3953"} Nov 24 02:27:09 crc kubenswrapper[4888]: I1124 02:27:09.538138 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-467dh/crc-debug-mkvxv" podStartSLOduration=2.315423666 podStartE2EDuration="13.538113414s" podCreationTimestamp="2025-11-24 02:26:56 +0000 UTC" firstStartedPulling="2025-11-24 02:26:57.217505277 +0000 UTC m=+7319.800189331" lastFinishedPulling="2025-11-24 02:27:08.440195045 +0000 UTC m=+7331.022879079" observedRunningTime="2025-11-24 02:27:09.527394413 +0000 UTC m=+7332.110078457" watchObservedRunningTime="2025-11-24 02:27:09.538113414 +0000 UTC m=+7332.120797458" Nov 24 02:28:04 crc kubenswrapper[4888]: I1124 02:28:04.065673 4888 generic.go:334] "Generic (PLEG): container finished" podID="bfebaba2-f323-4eb1-8bbc-93f71344b2ff" containerID="964a4ec403d0d3242e88140569203f1b5eae7cbb458a12fa63e802a3281a3953" exitCode=0 Nov 24 02:28:04 crc kubenswrapper[4888]: I1124 02:28:04.065746 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-mkvxv" event={"ID":"bfebaba2-f323-4eb1-8bbc-93f71344b2ff","Type":"ContainerDied","Data":"964a4ec403d0d3242e88140569203f1b5eae7cbb458a12fa63e802a3281a3953"} Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.232112 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.275528 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-467dh/crc-debug-mkvxv"] Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.286657 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-467dh/crc-debug-mkvxv"] Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.317173 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r\") pod \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.317396 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host\") pod \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\" (UID: \"bfebaba2-f323-4eb1-8bbc-93f71344b2ff\") " Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.317673 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host" (OuterVolumeSpecName: "host") pod "bfebaba2-f323-4eb1-8bbc-93f71344b2ff" (UID: "bfebaba2-f323-4eb1-8bbc-93f71344b2ff"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.318930 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.324431 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r" (OuterVolumeSpecName: "kube-api-access-5nq9r") pod "bfebaba2-f323-4eb1-8bbc-93f71344b2ff" (UID: "bfebaba2-f323-4eb1-8bbc-93f71344b2ff"). InnerVolumeSpecName "kube-api-access-5nq9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:28:05 crc kubenswrapper[4888]: I1124 02:28:05.421643 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/bfebaba2-f323-4eb1-8bbc-93f71344b2ff-kube-api-access-5nq9r\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.097808 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b86a7e425685a35e0e0f0a1119dd4e2473145016a236afa95d06a9827c0ef23" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.098505 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-mkvxv" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.271751 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfebaba2-f323-4eb1-8bbc-93f71344b2ff" path="/var/lib/kubelet/pods/bfebaba2-f323-4eb1-8bbc-93f71344b2ff/volumes" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.527129 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-467dh/crc-debug-vflwn"] Nov 24 02:28:06 crc kubenswrapper[4888]: E1124 02:28:06.527865 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfebaba2-f323-4eb1-8bbc-93f71344b2ff" containerName="container-00" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.527880 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfebaba2-f323-4eb1-8bbc-93f71344b2ff" containerName="container-00" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.528127 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfebaba2-f323-4eb1-8bbc-93f71344b2ff" containerName="container-00" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.528859 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.546110 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwvl9\" (UniqueName: \"kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.546168 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.649038 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwvl9\" (UniqueName: \"kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.649118 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.649277 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.682015 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwvl9\" (UniqueName: \"kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9\") pod \"crc-debug-vflwn\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:06 crc kubenswrapper[4888]: I1124 02:28:06.854052 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:07 crc kubenswrapper[4888]: I1124 02:28:07.111382 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-vflwn" event={"ID":"7ae528e3-b9eb-41e3-a566-636102033d38","Type":"ContainerStarted","Data":"94750f341e8b55ec1e6d3b7284519eb668d10e20e9eeb5a49c5a126ca44b79da"} Nov 24 02:28:08 crc kubenswrapper[4888]: I1124 02:28:08.124078 4888 generic.go:334] "Generic (PLEG): container finished" podID="7ae528e3-b9eb-41e3-a566-636102033d38" containerID="fbb272b1ab00299ae0dd66ee785314c74b368576c58d530c568119a702a9ec72" exitCode=0 Nov 24 02:28:08 crc kubenswrapper[4888]: I1124 02:28:08.124174 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-vflwn" event={"ID":"7ae528e3-b9eb-41e3-a566-636102033d38","Type":"ContainerDied","Data":"fbb272b1ab00299ae0dd66ee785314c74b368576c58d530c568119a702a9ec72"} Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.284194 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.311799 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host\") pod \"7ae528e3-b9eb-41e3-a566-636102033d38\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.311922 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwvl9\" (UniqueName: \"kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9\") pod \"7ae528e3-b9eb-41e3-a566-636102033d38\" (UID: \"7ae528e3-b9eb-41e3-a566-636102033d38\") " Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.311926 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host" (OuterVolumeSpecName: "host") pod "7ae528e3-b9eb-41e3-a566-636102033d38" (UID: "7ae528e3-b9eb-41e3-a566-636102033d38"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.315975 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae528e3-b9eb-41e3-a566-636102033d38-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.319971 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9" (OuterVolumeSpecName: "kube-api-access-rwvl9") pod "7ae528e3-b9eb-41e3-a566-636102033d38" (UID: "7ae528e3-b9eb-41e3-a566-636102033d38"). InnerVolumeSpecName "kube-api-access-rwvl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:28:09 crc kubenswrapper[4888]: I1124 02:28:09.421877 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwvl9\" (UniqueName: \"kubernetes.io/projected/7ae528e3-b9eb-41e3-a566-636102033d38-kube-api-access-rwvl9\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:10 crc kubenswrapper[4888]: I1124 02:28:10.147283 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-vflwn" event={"ID":"7ae528e3-b9eb-41e3-a566-636102033d38","Type":"ContainerDied","Data":"94750f341e8b55ec1e6d3b7284519eb668d10e20e9eeb5a49c5a126ca44b79da"} Nov 24 02:28:10 crc kubenswrapper[4888]: I1124 02:28:10.147549 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94750f341e8b55ec1e6d3b7284519eb668d10e20e9eeb5a49c5a126ca44b79da" Nov 24 02:28:10 crc kubenswrapper[4888]: I1124 02:28:10.147621 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-vflwn" Nov 24 02:28:10 crc kubenswrapper[4888]: I1124 02:28:10.530343 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-467dh/crc-debug-vflwn"] Nov 24 02:28:10 crc kubenswrapper[4888]: I1124 02:28:10.556727 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-467dh/crc-debug-vflwn"] Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.749721 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-467dh/crc-debug-lsjrx"] Nov 24 02:28:11 crc kubenswrapper[4888]: E1124 02:28:11.750228 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae528e3-b9eb-41e3-a566-636102033d38" containerName="container-00" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.750244 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae528e3-b9eb-41e3-a566-636102033d38" containerName="container-00" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.750555 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae528e3-b9eb-41e3-a566-636102033d38" containerName="container-00" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.751482 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.780910 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4fct\" (UniqueName: \"kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.781007 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.883383 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4fct\" (UniqueName: \"kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.883796 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.883939 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:11 crc kubenswrapper[4888]: I1124 02:28:11.902184 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4fct\" (UniqueName: \"kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct\") pod \"crc-debug-lsjrx\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:12 crc kubenswrapper[4888]: I1124 02:28:12.085139 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:12 crc kubenswrapper[4888]: I1124 02:28:12.173214 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-lsjrx" event={"ID":"ac702526-41cf-4f31-85c9-50a3c688efb6","Type":"ContainerStarted","Data":"609e634e3a645b309ba83f9899457deeb5b7fa6cf6e18618f4c22cdeddf5bc92"} Nov 24 02:28:12 crc kubenswrapper[4888]: I1124 02:28:12.270412 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae528e3-b9eb-41e3-a566-636102033d38" path="/var/lib/kubelet/pods/7ae528e3-b9eb-41e3-a566-636102033d38/volumes" Nov 24 02:28:13 crc kubenswrapper[4888]: I1124 02:28:13.188752 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/crc-debug-lsjrx" event={"ID":"ac702526-41cf-4f31-85c9-50a3c688efb6","Type":"ContainerDied","Data":"f0b2e307e0b72e2ea0f160912b66e7ee29d7bf56a1fefe05a4d34d47786da416"} Nov 24 02:28:13 crc kubenswrapper[4888]: I1124 02:28:13.189003 4888 generic.go:334] "Generic (PLEG): container finished" podID="ac702526-41cf-4f31-85c9-50a3c688efb6" containerID="f0b2e307e0b72e2ea0f160912b66e7ee29d7bf56a1fefe05a4d34d47786da416" exitCode=0 Nov 24 02:28:13 crc kubenswrapper[4888]: I1124 02:28:13.237691 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-467dh/crc-debug-lsjrx"] Nov 24 02:28:13 crc kubenswrapper[4888]: I1124 02:28:13.246928 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-467dh/crc-debug-lsjrx"] Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.312501 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.346442 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host\") pod \"ac702526-41cf-4f31-85c9-50a3c688efb6\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.346578 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host" (OuterVolumeSpecName: "host") pod "ac702526-41cf-4f31-85c9-50a3c688efb6" (UID: "ac702526-41cf-4f31-85c9-50a3c688efb6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.346619 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4fct\" (UniqueName: \"kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct\") pod \"ac702526-41cf-4f31-85c9-50a3c688efb6\" (UID: \"ac702526-41cf-4f31-85c9-50a3c688efb6\") " Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.347389 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac702526-41cf-4f31-85c9-50a3c688efb6-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.353085 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct" (OuterVolumeSpecName: "kube-api-access-g4fct") pod "ac702526-41cf-4f31-85c9-50a3c688efb6" (UID: "ac702526-41cf-4f31-85c9-50a3c688efb6"). InnerVolumeSpecName "kube-api-access-g4fct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:28:14 crc kubenswrapper[4888]: I1124 02:28:14.449967 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4fct\" (UniqueName: \"kubernetes.io/projected/ac702526-41cf-4f31-85c9-50a3c688efb6-kube-api-access-g4fct\") on node \"crc\" DevicePath \"\"" Nov 24 02:28:15 crc kubenswrapper[4888]: I1124 02:28:15.219450 4888 scope.go:117] "RemoveContainer" containerID="f0b2e307e0b72e2ea0f160912b66e7ee29d7bf56a1fefe05a4d34d47786da416" Nov 24 02:28:15 crc kubenswrapper[4888]: I1124 02:28:15.219550 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/crc-debug-lsjrx" Nov 24 02:28:16 crc kubenswrapper[4888]: I1124 02:28:16.271535 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac702526-41cf-4f31-85c9-50a3c688efb6" path="/var/lib/kubelet/pods/ac702526-41cf-4f31-85c9-50a3c688efb6/volumes" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.215529 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-api/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.410830 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-listener/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.442944 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-notifier/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.450971 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-evaluator/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.589659 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dcbbbf466-gt2vf_8e2c8c4b-730c-4c00-9dcd-03dd6b98b288/barbican-api/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.670769 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dcbbbf466-gt2vf_8e2c8c4b-730c-4c00-9dcd-03dd6b98b288/barbican-api-log/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.768042 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-96d8b7cb6-6rvkq_f68b1099-7b2b-4d27-961c-ac412f438292/barbican-keystone-listener/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.849469 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-847b7f6b9f-ngtnz_b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf/barbican-worker/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.916838 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-96d8b7cb6-6rvkq_f68b1099-7b2b-4d27-961c-ac412f438292/barbican-keystone-listener-log/0.log" Nov 24 02:28:54 crc kubenswrapper[4888]: I1124 02:28:54.981830 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-847b7f6b9f-ngtnz_b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf/barbican-worker-log/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.118303 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt_ebbc9282-a85b-4484-b130-5b0c91030009/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.257654 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/ceilometer-central-agent/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.285456 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/ceilometer-notification-agent/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.322777 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/proxy-httpd/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.396048 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/sg-core/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.502766 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv_ea89083c-a0ce-4bfe-b419-be89f9643924/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.632360 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds_99022481-7ff1-4b5a-a06e-376fbeaa29ba/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.741390 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6a2e8920-d187-4deb-85c8-b5b831af5995/cinder-api-log/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.829316 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6a2e8920-d187-4deb-85c8-b5b831af5995/cinder-api/0.log" Nov 24 02:28:55 crc kubenswrapper[4888]: I1124 02:28:55.980487 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5a08f210-f6e8-46f3-9796-8f4236a53dd1/probe/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.030368 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5a08f210-f6e8-46f3-9796-8f4236a53dd1/cinder-backup/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.157782 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_32d98628-e960-4e56-b620-7ba6050faadf/cinder-scheduler/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.258116 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_32d98628-e960-4e56-b620-7ba6050faadf/probe/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.366240 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a2644986-db7d-4368-956b-65dc26585e3b/cinder-volume/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.413798 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a2644986-db7d-4368-956b-65dc26585e3b/probe/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.569900 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-r99nj_5f1ab899-c7e5-4618-b282-8eebc0dec80b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.646144 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr_32bcbd1c-35f4-4632-ab42-01e6cc578d93/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:56 crc kubenswrapper[4888]: I1124 02:28:56.789684 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/init/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.041746 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/init/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.067262 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/dnsmasq-dns/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.077100 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5296f4e7-f819-43bb-9566-ee12ba6d51f0/glance-httpd/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.200150 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5296f4e7-f819-43bb-9566-ee12ba6d51f0/glance-log/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.264709 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2ae1990-d967-4fe3-80cf-9520e5b83286/glance-httpd/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.491405 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2ae1990-d967-4fe3-80cf-9520e5b83286/glance-log/0.log" Nov 24 02:28:57 crc kubenswrapper[4888]: I1124 02:28:57.929737 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-799c49dfcf-w6lnt_ef846e73-7a80-4d25-ac41-cfa95d68cc1e/heat-engine/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.284575 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7dfb67ff5b-2cglk_78ef5328-d894-4921-8464-62bdaf1c4208/horizon/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.494739 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv_7fce69b0-15d9-41d7-94f0-2263b74dff10/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.800347 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7dfb67ff5b-2cglk_78ef5328-d894-4921-8464-62bdaf1c4208/horizon-log/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.806545 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-rkgqz_e5f0ab5f-207b-4e2b-826a-87daf6aa5f25/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.824613 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-775b95987-88pxg_eb06d977-15f5-4263-aa7e-4fa2b2964847/heat-cfnapi/0.log" Nov 24 02:28:58 crc kubenswrapper[4888]: I1124 02:28:58.895410 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7f5f5cb65-zfpq5_6baa14da-0b8b-43da-81b9-605e596cbb11/heat-api/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.017133 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399101-qdc8s_81ab4fd4-2f1c-4023-b7a1-528e59bb56ad/keystone-cron/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.099011 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399161-vzqj4_5ab4f95d-24df-4a76-b678-a62601ae4375/keystone-cron/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.269871 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fe0c6981-4d50-4dba-b7a4-cabf7beb1947/kube-state-metrics/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.381948 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4_c4e89186-7454-4e0c-85ed-f7c90f6a7b0f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.442235 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b7f898bdd-bbfm7_b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10/keystone-api/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.540482 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-cvzdf_91206204-242c-4c3d-b686-9a4beebb76f1/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.662216 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8bcac458-0aa6-4a6c-8586-666803573def/manila-api-log/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.713196 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8bcac458-0aa6-4a6c-8586-666803573def/manila-api/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.837091 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eff408f2-f760-40a0-9016-b149c398cf85/probe/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.852045 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eff408f2-f760-40a0-9016-b149c398cf85/manila-scheduler/0.log" Nov 24 02:28:59 crc kubenswrapper[4888]: I1124 02:28:59.962718 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5f42ea1-62ff-4f74-bf41-4425b7332de8/manila-share/0.log" Nov 24 02:29:00 crc kubenswrapper[4888]: I1124 02:29:00.005281 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5f42ea1-62ff-4f74-bf41-4425b7332de8/probe/0.log" Nov 24 02:29:00 crc kubenswrapper[4888]: I1124 02:29:00.193266 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_f25aa6c8-4f87-4d83-8563-fc22a4bfa695/mysqld-exporter/0.log" Nov 24 02:29:00 crc kubenswrapper[4888]: I1124 02:29:00.572915 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv_fb3242a5-393f-4fda-a791-0d1246a5bded/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:00 crc kubenswrapper[4888]: I1124 02:29:00.623361 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-686d6d8cd9-m24dl_d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16/neutron-api/0.log" Nov 24 02:29:00 crc kubenswrapper[4888]: I1124 02:29:00.649904 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-686d6d8cd9-m24dl_d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16/neutron-httpd/0.log" Nov 24 02:29:01 crc kubenswrapper[4888]: I1124 02:29:01.109880 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_be8f5419-cba3-4446-937d-f8d4b06c1c0e/nova-cell0-conductor-conductor/0.log" Nov 24 02:29:01 crc kubenswrapper[4888]: I1124 02:29:01.395759 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f5cc70d4-9318-4552-825b-9ad495d378b0/nova-cell1-conductor-conductor/0.log" Nov 24 02:29:01 crc kubenswrapper[4888]: I1124 02:29:01.505996 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f4060c08-c47b-483a-8a17-661a171187f2/nova-api-log/0.log" Nov 24 02:29:01 crc kubenswrapper[4888]: I1124 02:29:01.739194 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a5e4972d-5b76-4f46-a9fa-3a4425534ad8/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 02:29:01 crc kubenswrapper[4888]: I1124 02:29:01.858228 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt_59680387-8e94-4e51-9b44-9c0829e262a6/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.018674 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfb6aad7-f065-42f1-9021-f5debe134c7d/nova-metadata-log/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.048717 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f4060c08-c47b-483a-8a17-661a171187f2/nova-api-api/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.461797 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/mysql-bootstrap/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.471182 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_894a3216-d75f-4229-849f-4b7f4e942d96/nova-scheduler-scheduler/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.672427 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/mysql-bootstrap/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.678217 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/galera/0.log" Nov 24 02:29:02 crc kubenswrapper[4888]: I1124 02:29:02.919274 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/mysql-bootstrap/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.194074 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/galera/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.199528 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/mysql-bootstrap/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.421553 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a036e658-146d-4eb0-a197-939e6bbc31f8/openstackclient/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.426313 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4s282_6f28f199-4a0d-4c00-8aef-c5219f6fcc10/openstack-network-exporter/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.684397 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server-init/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.840491 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server-init/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.940629 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server/0.log" Nov 24 02:29:03 crc kubenswrapper[4888]: I1124 02:29:03.943877 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovs-vswitchd/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.142530 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x5tvh_053aa355-b781-49f4-8e8a-80ae64c4de62/ovn-controller/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.358179 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gjwnp_26532baf-63c3-4d2f-87c2-9c690c53823c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.419691 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c37246a-89a8-4f28-95f9-149b6c9f1790/openstack-network-exporter/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.562260 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c37246a-89a8-4f28-95f9-149b6c9f1790/ovn-northd/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.660651 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_728a0118-33ee-4fb3-b8e2-198e075adc31/openstack-network-exporter/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.816955 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_728a0118-33ee-4fb3-b8e2-198e075adc31/ovsdbserver-nb/0.log" Nov 24 02:29:04 crc kubenswrapper[4888]: I1124 02:29:04.907245 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cdce96a2-6d42-4761-9170-ef61211cfc92/openstack-network-exporter/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.061318 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cdce96a2-6d42-4761-9170-ef61211cfc92/ovsdbserver-sb/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.113634 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfb6aad7-f065-42f1-9021-f5debe134c7d/nova-metadata-metadata/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.275968 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-669f454766-khh4h_4c22d012-f519-4556-887a-2a0a11d7f3ec/placement-api/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.421224 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/init-config-reloader/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.439468 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-669f454766-khh4h_4c22d012-f519-4556-887a-2a0a11d7f3ec/placement-log/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.615408 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/init-config-reloader/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.621296 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/config-reloader/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.669748 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/thanos-sidecar/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.680424 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/prometheus/0.log" Nov 24 02:29:05 crc kubenswrapper[4888]: I1124 02:29:05.840272 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/setup-container/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.059100 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/setup-container/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.079582 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/setup-container/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.194802 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/rabbitmq/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.403238 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/setup-container/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.429333 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/rabbitmq/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.445708 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll_2d25d7f9-02eb-4f78-a3c9-2f9a9a533797/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.664867 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt_30e410dc-697f-4714-9e4b-a2f8a71796c4/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.757689 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jrjs2_bab960db-54de-4def-b56e-9f31b8df0e03/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:06 crc kubenswrapper[4888]: I1124 02:29:06.926976 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pbl78_8804f688-d740-4f33-ba57-ac0ca3d63a2b/ssh-known-hosts-edpm-deployment/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.101727 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6dd96dd9bc-c9x6n_07c3d65f-a35c-481d-a773-a9dc4dd914b5/proxy-server/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.269262 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bq8p7_1eaecd0e-0a6f-45d1-95da-c6ed86192911/swift-ring-rebalance/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.314030 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6dd96dd9bc-c9x6n_07c3d65f-a35c-481d-a773-a9dc4dd914b5/proxy-httpd/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.423949 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-auditor/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.511182 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-reaper/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.597313 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-server/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.617588 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-replicator/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.656993 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-auditor/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.757830 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-replicator/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.817423 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-updater/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.837264 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-server/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.916572 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-auditor/0.log" Nov 24 02:29:07 crc kubenswrapper[4888]: I1124 02:29:07.969373 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-expirer/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.062204 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-server/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.098191 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-replicator/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.191678 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-updater/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.232761 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/swift-recon-cron/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.256369 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/rsync/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.463251 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-t54vc_7009799b-7622-43c3-856d-7a368de7608f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.574446 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9_1fb69384-3ad1-424b-93ad-2716c502bc4d/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:08 crc kubenswrapper[4888]: I1124 02:29:08.796568 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717/test-operator-logs-container/0.log" Nov 24 02:29:09 crc kubenswrapper[4888]: I1124 02:29:09.008938 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-crwxl_de6be747-eec3-445b-a6e5-3cd034b68c6a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:29:09 crc kubenswrapper[4888]: I1124 02:29:09.651114 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cc931795-8f17-4afd-a16c-264855b1755c/tempest-tests-tempest-tests-runner/0.log" Nov 24 02:29:10 crc kubenswrapper[4888]: I1124 02:29:10.952071 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:10 crc kubenswrapper[4888]: E1124 02:29:10.952785 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac702526-41cf-4f31-85c9-50a3c688efb6" containerName="container-00" Nov 24 02:29:10 crc kubenswrapper[4888]: I1124 02:29:10.952800 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac702526-41cf-4f31-85c9-50a3c688efb6" containerName="container-00" Nov 24 02:29:10 crc kubenswrapper[4888]: I1124 02:29:10.953036 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac702526-41cf-4f31-85c9-50a3c688efb6" containerName="container-00" Nov 24 02:29:10 crc kubenswrapper[4888]: I1124 02:29:10.954563 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.059282 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.072233 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jgmd\" (UniqueName: \"kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.072327 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.072354 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.196853 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jgmd\" (UniqueName: \"kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.197035 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.197090 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.198193 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.198484 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.236795 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jgmd\" (UniqueName: \"kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd\") pod \"redhat-operators-ps2jk\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:11 crc kubenswrapper[4888]: I1124 02:29:11.295112 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:12 crc kubenswrapper[4888]: I1124 02:29:12.407388 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:12 crc kubenswrapper[4888]: I1124 02:29:12.876706 4888 generic.go:334] "Generic (PLEG): container finished" podID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerID="4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d" exitCode=0 Nov 24 02:29:12 crc kubenswrapper[4888]: I1124 02:29:12.876750 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerDied","Data":"4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d"} Nov 24 02:29:12 crc kubenswrapper[4888]: I1124 02:29:12.876777 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerStarted","Data":"2e96f3cc9b5bf0c9fa339716e200c239ec627646d34713ad774538a94617a51a"} Nov 24 02:29:12 crc kubenswrapper[4888]: I1124 02:29:12.878980 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:29:13 crc kubenswrapper[4888]: I1124 02:29:13.905318 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerStarted","Data":"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460"} Nov 24 02:29:17 crc kubenswrapper[4888]: I1124 02:29:17.956755 4888 generic.go:334] "Generic (PLEG): container finished" podID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerID="d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460" exitCode=0 Nov 24 02:29:17 crc kubenswrapper[4888]: I1124 02:29:17.956841 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerDied","Data":"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460"} Nov 24 02:29:18 crc kubenswrapper[4888]: I1124 02:29:18.968436 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerStarted","Data":"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e"} Nov 24 02:29:18 crc kubenswrapper[4888]: I1124 02:29:18.984230 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ps2jk" podStartSLOduration=3.500966848 podStartE2EDuration="8.984215497s" podCreationTimestamp="2025-11-24 02:29:10 +0000 UTC" firstStartedPulling="2025-11-24 02:29:12.878723519 +0000 UTC m=+7455.461407563" lastFinishedPulling="2025-11-24 02:29:18.361972168 +0000 UTC m=+7460.944656212" observedRunningTime="2025-11-24 02:29:18.983307901 +0000 UTC m=+7461.565991945" watchObservedRunningTime="2025-11-24 02:29:18.984215497 +0000 UTC m=+7461.566899541" Nov 24 02:29:21 crc kubenswrapper[4888]: I1124 02:29:21.295311 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:21 crc kubenswrapper[4888]: I1124 02:29:21.296328 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:22 crc kubenswrapper[4888]: I1124 02:29:22.347972 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ps2jk" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" probeResult="failure" output=< Nov 24 02:29:22 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:29:22 crc kubenswrapper[4888]: > Nov 24 02:29:23 crc kubenswrapper[4888]: I1124 02:29:23.591145 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:29:23 crc kubenswrapper[4888]: I1124 02:29:23.592156 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:29:24 crc kubenswrapper[4888]: I1124 02:29:24.310507 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fca78ec3-fec9-4c3e-aee9-75d3e049c668/memcached/0.log" Nov 24 02:29:32 crc kubenswrapper[4888]: I1124 02:29:32.349392 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ps2jk" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" probeResult="failure" output=< Nov 24 02:29:32 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:29:32 crc kubenswrapper[4888]: > Nov 24 02:29:39 crc kubenswrapper[4888]: I1124 02:29:39.666636 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:29:39 crc kubenswrapper[4888]: I1124 02:29:39.813895 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:29:39 crc kubenswrapper[4888]: I1124 02:29:39.859251 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:29:39 crc kubenswrapper[4888]: I1124 02:29:39.859445 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.044612 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.064040 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/extract/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.066596 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.232792 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-gphtr_e7342ffc-1f02-4eca-9be9-5816f5f0c86c/kube-rbac-proxy/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.302550 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-gphtr_e7342ffc-1f02-4eca-9be9-5816f5f0c86c/manager/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.322223 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-4gh28_5fe80165-08e7-4a34-8290-fd0d7fd22dd2/kube-rbac-proxy/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.491167 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-4gh28_5fe80165-08e7-4a34-8290-fd0d7fd22dd2/manager/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.511057 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-9nb4l_a277e2cc-4990-4d79-8a4f-ec03c81d747d/kube-rbac-proxy/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.567269 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-9nb4l_a277e2cc-4990-4d79-8a4f-ec03c81d747d/manager/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.715127 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-mtqr5_a701224c-a2ad-49b6-98b7-55fd37051e5b/kube-rbac-proxy/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.806703 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-mtqr5_a701224c-a2ad-49b6-98b7-55fd37051e5b/manager/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.902738 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-wbxrh_6cf992a1-eac0-4269-b8bb-e14554116727/kube-rbac-proxy/0.log" Nov 24 02:29:40 crc kubenswrapper[4888]: I1124 02:29:40.987923 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-wbxrh_6cf992a1-eac0-4269-b8bb-e14554116727/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.050985 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-llt2h_10a23051-0913-44fd-abda-89436a7e3b59/kube-rbac-proxy/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.104356 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-llt2h_10a23051-0913-44fd-abda-89436a7e3b59/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.174215 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-zl954_e6420e55-4951-4546-b252-44c25d49a0c3/kube-rbac-proxy/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.397076 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jk6ph_3766dfec-2de1-4df2-8609-0cbbe8599342/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.434077 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jk6ph_3766dfec-2de1-4df2-8609-0cbbe8599342/kube-rbac-proxy/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.440533 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-zl954_e6420e55-4951-4546-b252-44c25d49a0c3/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.618742 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-8kqtd_587fee04-3081-4c59-8125-0e47436e2347/kube-rbac-proxy/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.683554 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-8kqtd_587fee04-3081-4c59-8125-0e47436e2347/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.777324 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-tsb9j_c18797ad-a3dc-46af-9aaf-d9e17918228a/kube-rbac-proxy/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.860664 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-tsb9j_c18797ad-a3dc-46af-9aaf-d9e17918228a/manager/0.log" Nov 24 02:29:41 crc kubenswrapper[4888]: I1124 02:29:41.876838 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-4pmdz_dbaba630-daf7-46c3-807a-6e76128fc7d6/kube-rbac-proxy/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.051138 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-4pmdz_dbaba630-daf7-46c3-807a-6e76128fc7d6/manager/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.136788 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-fn4cs_afa14c5d-d820-4b05-bddd-4d00e49a421a/kube-rbac-proxy/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.200901 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-fn4cs_afa14c5d-d820-4b05-bddd-4d00e49a421a/manager/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.314569 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-7zjtv_f4bc41bd-b9d5-4acf-978e-75144d8b43a3/kube-rbac-proxy/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.352021 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ps2jk" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" probeResult="failure" output=< Nov 24 02:29:42 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:29:42 crc kubenswrapper[4888]: > Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.452356 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-7zjtv_f4bc41bd-b9d5-4acf-978e-75144d8b43a3/manager/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.521301 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-vxffn_750c449f-a775-4ee8-82b8-a6be531250fe/manager/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.548299 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-vxffn_750c449f-a775-4ee8-82b8-a6be531250fe/kube-rbac-proxy/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.698444 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh_2b35fb96-9ea1-4aaa-9396-cd5f705f63bb/kube-rbac-proxy/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.748715 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh_2b35fb96-9ea1-4aaa-9396-cd5f705f63bb/manager/0.log" Nov 24 02:29:42 crc kubenswrapper[4888]: I1124 02:29:42.912925 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d84888d56-jk5d2_db75890a-4c13-4f98-a410-aa4f98890ccc/kube-rbac-proxy/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.015314 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75b7b5b9df-7j8cf_9cc98d14-8e44-4556-b00a-ed2069d395cf/kube-rbac-proxy/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.247749 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ggr8j_927e1330-0a2c-4e33-b97d-5de99ce8138c/registry-server/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.375061 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75b7b5b9df-7j8cf_9cc98d14-8e44-4556-b00a-ed2069d395cf/operator/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.505352 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-c9cfn_8a57fb36-c0db-4778-9c17-7241b5e8f602/kube-rbac-proxy/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.622061 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-c9cfn_8a57fb36-c0db-4778-9c17-7241b5e8f602/manager/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.709496 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-98b6n_fef147fd-ee85-4c05-9652-5e44223a3e5e/kube-rbac-proxy/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.743825 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-98b6n_fef147fd-ee85-4c05-9652-5e44223a3e5e/manager/0.log" Nov 24 02:29:43 crc kubenswrapper[4888]: I1124 02:29:43.955169 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv_e569ce1f-9272-4c6a-bb91-a01616bb5883/operator/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.009158 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-5ck2h_3a5b66a7-1cf6-4840-870c-f7a6968bd08a/kube-rbac-proxy/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.177222 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67c4db9997-7rqlj_bc56a88d-e719-40f5-a210-1b7afdffae8b/kube-rbac-proxy/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.224670 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-5ck2h_3a5b66a7-1cf6-4840-870c-f7a6968bd08a/manager/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.410946 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-mn678_82f2609b-e676-49f8-9ab4-d06fb3e626be/kube-rbac-proxy/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.509219 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-mn678_82f2609b-e676-49f8-9ab4-d06fb3e626be/manager/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.673658 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67c4db9997-7rqlj_bc56a88d-e719-40f5-a210-1b7afdffae8b/manager/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.712385 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d84888d56-jk5d2_db75890a-4c13-4f98-a410-aa4f98890ccc/manager/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.720486 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-6mmtv_c36e3742-c48a-48d1-96c8-87c3809a4183/manager/0.log" Nov 24 02:29:44 crc kubenswrapper[4888]: I1124 02:29:44.742587 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-6mmtv_c36e3742-c48a-48d1-96c8-87c3809a4183/kube-rbac-proxy/0.log" Nov 24 02:29:51 crc kubenswrapper[4888]: I1124 02:29:51.349053 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:51 crc kubenswrapper[4888]: I1124 02:29:51.415335 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:51 crc kubenswrapper[4888]: I1124 02:29:51.598168 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:52 crc kubenswrapper[4888]: I1124 02:29:52.530573 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ps2jk" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" containerID="cri-o://6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e" gracePeriod=2 Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.145027 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.294493 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content\") pod \"f1adb7e9-4202-4475-8028-7bd47fefd071\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.294640 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities\") pod \"f1adb7e9-4202-4475-8028-7bd47fefd071\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.294881 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jgmd\" (UniqueName: \"kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd\") pod \"f1adb7e9-4202-4475-8028-7bd47fefd071\" (UID: \"f1adb7e9-4202-4475-8028-7bd47fefd071\") " Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.295405 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities" (OuterVolumeSpecName: "utilities") pod "f1adb7e9-4202-4475-8028-7bd47fefd071" (UID: "f1adb7e9-4202-4475-8028-7bd47fefd071"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.308089 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd" (OuterVolumeSpecName: "kube-api-access-5jgmd") pod "f1adb7e9-4202-4475-8028-7bd47fefd071" (UID: "f1adb7e9-4202-4475-8028-7bd47fefd071"). InnerVolumeSpecName "kube-api-access-5jgmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.399415 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jgmd\" (UniqueName: \"kubernetes.io/projected/f1adb7e9-4202-4475-8028-7bd47fefd071-kube-api-access-5jgmd\") on node \"crc\" DevicePath \"\"" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.399454 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.414503 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1adb7e9-4202-4475-8028-7bd47fefd071" (UID: "f1adb7e9-4202-4475-8028-7bd47fefd071"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.501597 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1adb7e9-4202-4475-8028-7bd47fefd071-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.542428 4888 generic.go:334] "Generic (PLEG): container finished" podID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerID="6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e" exitCode=0 Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.542497 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerDied","Data":"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e"} Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.542521 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ps2jk" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.542542 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ps2jk" event={"ID":"f1adb7e9-4202-4475-8028-7bd47fefd071","Type":"ContainerDied","Data":"2e96f3cc9b5bf0c9fa339716e200c239ec627646d34713ad774538a94617a51a"} Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.542566 4888 scope.go:117] "RemoveContainer" containerID="6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.569772 4888 scope.go:117] "RemoveContainer" containerID="d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.576970 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.588183 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ps2jk"] Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.591374 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.591414 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.601503 4888 scope.go:117] "RemoveContainer" containerID="4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.647684 4888 scope.go:117] "RemoveContainer" containerID="6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e" Nov 24 02:29:53 crc kubenswrapper[4888]: E1124 02:29:53.648243 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e\": container with ID starting with 6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e not found: ID does not exist" containerID="6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.648295 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e"} err="failed to get container status \"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e\": rpc error: code = NotFound desc = could not find container \"6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e\": container with ID starting with 6b50d57571bed3762790a5b677ee10f71fa123d156359d1188baf80608bef27e not found: ID does not exist" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.648327 4888 scope.go:117] "RemoveContainer" containerID="d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460" Nov 24 02:29:53 crc kubenswrapper[4888]: E1124 02:29:53.649197 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460\": container with ID starting with d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460 not found: ID does not exist" containerID="d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.649228 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460"} err="failed to get container status \"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460\": rpc error: code = NotFound desc = could not find container \"d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460\": container with ID starting with d21bb6a8474bf0b9bcc8af12cf245645f5421355890c4b342ec71b5a4e2a4460 not found: ID does not exist" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.649251 4888 scope.go:117] "RemoveContainer" containerID="4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d" Nov 24 02:29:53 crc kubenswrapper[4888]: E1124 02:29:53.649511 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d\": container with ID starting with 4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d not found: ID does not exist" containerID="4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d" Nov 24 02:29:53 crc kubenswrapper[4888]: I1124 02:29:53.649548 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d"} err="failed to get container status \"4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d\": rpc error: code = NotFound desc = could not find container \"4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d\": container with ID starting with 4c7cd2b89d6f6f153cd5709da622a4a75c470cc65efadc1fbe4d464e2603d89d not found: ID does not exist" Nov 24 02:29:54 crc kubenswrapper[4888]: I1124 02:29:54.256868 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" path="/var/lib/kubelet/pods/f1adb7e9-4202-4475-8028-7bd47fefd071/volumes" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.159643 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc"] Nov 24 02:30:00 crc kubenswrapper[4888]: E1124 02:30:00.161480 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="extract-utilities" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.161520 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="extract-utilities" Nov 24 02:30:00 crc kubenswrapper[4888]: E1124 02:30:00.161540 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="extract-content" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.161549 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="extract-content" Nov 24 02:30:00 crc kubenswrapper[4888]: E1124 02:30:00.161589 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.161597 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.161911 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1adb7e9-4202-4475-8028-7bd47fefd071" containerName="registry-server" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.162893 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.169031 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.169022 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.171966 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc"] Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.257281 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjn7w\" (UniqueName: \"kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.257444 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.257517 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.359611 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.359699 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.359981 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjn7w\" (UniqueName: \"kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.362131 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.368459 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.380377 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjn7w\" (UniqueName: \"kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w\") pod \"collect-profiles-29399190-mtlhc\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:00 crc kubenswrapper[4888]: I1124 02:30:00.488293 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.010404 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc"] Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.616212 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j6829_814a1313-1c15-436c-9a1d-132859fabdea/control-plane-machine-set-operator/0.log" Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.634768 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" event={"ID":"a786868b-d10c-4169-804a-3f25038a30cf","Type":"ContainerStarted","Data":"05b98a6b98215828f3799c22f8f7bbd492518872e8e07a2657aded707c08157f"} Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.634839 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" event={"ID":"a786868b-d10c-4169-804a-3f25038a30cf","Type":"ContainerStarted","Data":"e2f4d1c19ed6e8b6b35a71452a0ae0e7e58b1c25ddbefa1b25a9ba3ecbeed85e"} Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.662042 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" podStartSLOduration=1.6620241839999998 podStartE2EDuration="1.662024184s" podCreationTimestamp="2025-11-24 02:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:30:01.655508971 +0000 UTC m=+7504.238193025" watchObservedRunningTime="2025-11-24 02:30:01.662024184 +0000 UTC m=+7504.244708228" Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.823894 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8s79_103c3930-7610-4c38-af13-cb0f06da2629/kube-rbac-proxy/0.log" Nov 24 02:30:01 crc kubenswrapper[4888]: I1124 02:30:01.848072 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8s79_103c3930-7610-4c38-af13-cb0f06da2629/machine-api-operator/0.log" Nov 24 02:30:02 crc kubenswrapper[4888]: I1124 02:30:02.646611 4888 generic.go:334] "Generic (PLEG): container finished" podID="a786868b-d10c-4169-804a-3f25038a30cf" containerID="05b98a6b98215828f3799c22f8f7bbd492518872e8e07a2657aded707c08157f" exitCode=0 Nov 24 02:30:02 crc kubenswrapper[4888]: I1124 02:30:02.646667 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" event={"ID":"a786868b-d10c-4169-804a-3f25038a30cf","Type":"ContainerDied","Data":"05b98a6b98215828f3799c22f8f7bbd492518872e8e07a2657aded707c08157f"} Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.209427 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.356564 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjn7w\" (UniqueName: \"kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w\") pod \"a786868b-d10c-4169-804a-3f25038a30cf\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.356745 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume\") pod \"a786868b-d10c-4169-804a-3f25038a30cf\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.357002 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume\") pod \"a786868b-d10c-4169-804a-3f25038a30cf\" (UID: \"a786868b-d10c-4169-804a-3f25038a30cf\") " Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.357777 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume" (OuterVolumeSpecName: "config-volume") pod "a786868b-d10c-4169-804a-3f25038a30cf" (UID: "a786868b-d10c-4169-804a-3f25038a30cf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.364894 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a786868b-d10c-4169-804a-3f25038a30cf" (UID: "a786868b-d10c-4169-804a-3f25038a30cf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.364959 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w" (OuterVolumeSpecName: "kube-api-access-gjn7w") pod "a786868b-d10c-4169-804a-3f25038a30cf" (UID: "a786868b-d10c-4169-804a-3f25038a30cf"). InnerVolumeSpecName "kube-api-access-gjn7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.459609 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjn7w\" (UniqueName: \"kubernetes.io/projected/a786868b-d10c-4169-804a-3f25038a30cf-kube-api-access-gjn7w\") on node \"crc\" DevicePath \"\"" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.459853 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a786868b-d10c-4169-804a-3f25038a30cf-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.459959 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a786868b-d10c-4169-804a-3f25038a30cf-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.672940 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" event={"ID":"a786868b-d10c-4169-804a-3f25038a30cf","Type":"ContainerDied","Data":"e2f4d1c19ed6e8b6b35a71452a0ae0e7e58b1c25ddbefa1b25a9ba3ecbeed85e"} Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.673270 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f4d1c19ed6e8b6b35a71452a0ae0e7e58b1c25ddbefa1b25a9ba3ecbeed85e" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.673018 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399190-mtlhc" Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.731210 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2"] Nov 24 02:30:04 crc kubenswrapper[4888]: I1124 02:30:04.741576 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399145-5s2t2"] Nov 24 02:30:06 crc kubenswrapper[4888]: I1124 02:30:06.261642 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9" path="/var/lib/kubelet/pods/1bc1aff7-8c2a-4f9a-bc9e-c56d72cf19b9/volumes" Nov 24 02:30:15 crc kubenswrapper[4888]: I1124 02:30:15.164727 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2ls5n_087d3537-623e-4f8b-be5b-b9342a2cf19f/cert-manager-controller/0.log" Nov 24 02:30:15 crc kubenswrapper[4888]: I1124 02:30:15.324544 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vjmm8_25c1c650-be40-437b-9ea3-8e9a60b2f2f2/cert-manager-cainjector/0.log" Nov 24 02:30:15 crc kubenswrapper[4888]: I1124 02:30:15.366568 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-g8kdr_6a61128e-aa42-4201-87ae-856cc4e6f682/cert-manager-webhook/0.log" Nov 24 02:30:22 crc kubenswrapper[4888]: I1124 02:30:22.803384 4888 scope.go:117] "RemoveContainer" containerID="64504b20012dd03a180271541054768948565107bcd6881b8077bbe1a2b4159f" Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.591597 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.592134 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.592194 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.592714 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.592764 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58" gracePeriod=600 Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.870332 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58" exitCode=0 Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.870396 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58"} Nov 24 02:30:23 crc kubenswrapper[4888]: I1124 02:30:23.870726 4888 scope.go:117] "RemoveContainer" containerID="aee1e9d1ef6867382e49cb9a4976d0ce41582760665fd7715bbc6857e0535de6" Nov 24 02:30:24 crc kubenswrapper[4888]: I1124 02:30:24.887904 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec"} Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.103833 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-vzwm2_d88df599-7094-4325-afeb-d4188a9b948c/nmstate-console-plugin/0.log" Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.251219 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fdwl9_00daa12e-54c4-46e9-835c-0a2d2c9ad90e/nmstate-handler/0.log" Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.348286 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kgwmb_f9bc22c3-35a5-4afc-90fe-67b9c086b46a/kube-rbac-proxy/0.log" Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.396511 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kgwmb_f9bc22c3-35a5-4afc-90fe-67b9c086b46a/nmstate-metrics/0.log" Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.511783 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-mncvc_f1648233-9c9f-4a7a-8acc-35d91c1a0b04/nmstate-operator/0.log" Nov 24 02:30:28 crc kubenswrapper[4888]: I1124 02:30:28.584242 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-fhrr5_2737508c-e47e-4ae4-b412-b665096fbf78/nmstate-webhook/0.log" Nov 24 02:30:41 crc kubenswrapper[4888]: I1124 02:30:41.081709 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/manager/0.log" Nov 24 02:30:41 crc kubenswrapper[4888]: I1124 02:30:41.100331 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/kube-rbac-proxy/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.421172 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lv57g_2b99942b-b619-4a73-bae0-b3cd34990e42/cluster-logging-operator/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.548881 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-gmbz6_61cf37cb-90a9-4b6a-8252-30a564f91b0d/collector/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.596411 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_38b36998-57a3-4b56-b3dd-16f2800890d8/loki-compactor/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.739091 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-4kw2j_9dfccf91-b0dd-46c1-a8a0-784d83237bbe/loki-distributor/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.784095 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-7m9r2_7000141c-3ebe-45cf-bdc1-36be274b4d8c/gateway/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.802635 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-7m9r2_7000141c-3ebe-45cf-bdc1-36be274b4d8c/opa/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.953308 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-hjsdr_8c68d551-242e-47b9-8ed7-d47f6f256d3f/opa/0.log" Nov 24 02:30:55 crc kubenswrapper[4888]: I1124 02:30:55.966715 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-hjsdr_8c68d551-242e-47b9-8ed7-d47f6f256d3f/gateway/0.log" Nov 24 02:30:56 crc kubenswrapper[4888]: I1124 02:30:56.072648 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_e8d0d952-6592-40e8-a831-5e875e1f9d4b/loki-index-gateway/0.log" Nov 24 02:30:56 crc kubenswrapper[4888]: I1124 02:30:56.188454 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_158da35a-1b60-4ef6-b46a-fd8d3e8c5c51/loki-ingester/0.log" Nov 24 02:30:56 crc kubenswrapper[4888]: I1124 02:30:56.237404 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-69xbm_03bd362f-47ff-4c5c-8ae8-b60c0c87138b/loki-querier/0.log" Nov 24 02:30:56 crc kubenswrapper[4888]: I1124 02:30:56.395670 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-ntcth_2386f5e1-9382-4d87-b641-4372f630d9d4/loki-query-frontend/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.087729 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-77rp4_e710f559-91d6-47ce-bf7d-ee5717c276a8/kube-rbac-proxy/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.264944 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-77rp4_e710f559-91d6-47ce-bf7d-ee5717c276a8/controller/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.336472 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.553649 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.571382 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.571494 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.572949 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.786063 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.786966 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.809782 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.833859 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:31:10 crc kubenswrapper[4888]: I1124 02:31:10.970172 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.004983 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.010471 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/controller/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.049274 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.189045 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/frr-metrics/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.205351 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/kube-rbac-proxy/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.282485 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/kube-rbac-proxy-frr/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.425833 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/reloader/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.532986 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-zvfv2_43445b39-e7aa-4d6c-bc1a-760e001605b3/frr-k8s-webhook-server/0.log" Nov 24 02:31:11 crc kubenswrapper[4888]: I1124 02:31:11.962095 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f89595f9b-rs4gm_d8651fd8-33fc-4d15-b478-78921c04fef1/manager/0.log" Nov 24 02:31:12 crc kubenswrapper[4888]: I1124 02:31:12.116893 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69f6b76c88-dh8sn_007267a1-8af0-47c2-84e1-8fa533e3f5c7/webhook-server/0.log" Nov 24 02:31:12 crc kubenswrapper[4888]: I1124 02:31:12.161025 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2lxkv_2b50d2f6-eedb-4817-ba29-4b6e908aae96/kube-rbac-proxy/0.log" Nov 24 02:31:12 crc kubenswrapper[4888]: I1124 02:31:12.873764 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2lxkv_2b50d2f6-eedb-4817-ba29-4b6e908aae96/speaker/0.log" Nov 24 02:31:13 crc kubenswrapper[4888]: I1124 02:31:13.117322 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/frr/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.444922 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.620698 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.627084 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.635421 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.786563 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.810348 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/extract/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.817462 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:31:25 crc kubenswrapper[4888]: I1124 02:31:25.955334 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.103951 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.128746 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.129570 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.317056 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.327892 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/extract/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.333279 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.516524 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.655034 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.707713 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.718905 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.908898 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.920285 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/extract/0.log" Nov 24 02:31:26 crc kubenswrapper[4888]: I1124 02:31:26.921860 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.082239 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.278126 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.281369 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.290235 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.441982 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.443522 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/extract/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.466572 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.596850 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.786859 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.804919 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:31:27 crc kubenswrapper[4888]: I1124 02:31:27.828667 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.012458 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.039895 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.222804 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.515100 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.516633 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.553004 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.796802 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:31:28 crc kubenswrapper[4888]: I1124 02:31:28.802708 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.078232 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.131866 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/registry-server/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.182838 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.220381 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.265689 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.454880 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.455877 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.505313 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/extract/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.684590 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nmmd_ef22eb11-250d-44e2-a2bf-7906d7817b8b/marketplace-operator/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.743833 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.864490 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/registry-server/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.935833 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.940590 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:31:29 crc kubenswrapper[4888]: I1124 02:31:29.941438 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.097967 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.098665 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.164889 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.323239 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/registry-server/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.370047 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.377107 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.424026 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.596833 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:31:30 crc kubenswrapper[4888]: I1124 02:31:30.609543 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:31:31 crc kubenswrapper[4888]: I1124 02:31:31.364946 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/registry-server/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.086115 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-x756x_85dc9246-ac2d-44f7-bda0-bdd6faa6a391/prometheus-operator/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.238085 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_a11b0ce1-5305-4b79-a302-4839c19791e0/prometheus-operator-admission-webhook/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.295504 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4/prometheus-operator-admission-webhook/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.421379 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-vj7xs_2c9cc7d0-818c-484c-96fd-c22c437348bc/operator/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.483631 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-clzxk_a05abc8e-614d-4cb7-8eda-35f0698a756c/observability-ui-dashboards/0.log" Nov 24 02:31:43 crc kubenswrapper[4888]: I1124 02:31:43.577499 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-7j7tc_920040ee-4308-431b-9f6d-eb16563acb3f/perses-operator/0.log" Nov 24 02:31:56 crc kubenswrapper[4888]: I1124 02:31:56.235456 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/kube-rbac-proxy/0.log" Nov 24 02:31:56 crc kubenswrapper[4888]: I1124 02:31:56.271957 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/manager/0.log" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.527315 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:04 crc kubenswrapper[4888]: E1124 02:32:04.529998 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a786868b-d10c-4169-804a-3f25038a30cf" containerName="collect-profiles" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.530118 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="a786868b-d10c-4169-804a-3f25038a30cf" containerName="collect-profiles" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.530492 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="a786868b-d10c-4169-804a-3f25038a30cf" containerName="collect-profiles" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.532893 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.558753 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.623657 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.624004 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.624420 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj8rp\" (UniqueName: \"kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.726862 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.726997 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.727167 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj8rp\" (UniqueName: \"kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.727471 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.727473 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.758079 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj8rp\" (UniqueName: \"kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp\") pod \"community-operators-2gkxr\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:04 crc kubenswrapper[4888]: I1124 02:32:04.875525 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:05 crc kubenswrapper[4888]: I1124 02:32:05.313364 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:05 crc kubenswrapper[4888]: I1124 02:32:05.982831 4888 generic.go:334] "Generic (PLEG): container finished" podID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerID="668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267" exitCode=0 Nov 24 02:32:05 crc kubenswrapper[4888]: I1124 02:32:05.982924 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerDied","Data":"668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267"} Nov 24 02:32:05 crc kubenswrapper[4888]: I1124 02:32:05.983118 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerStarted","Data":"08570674a171a77b5b1d6ec50c3794bae2f840d1f17ca1db7995bb609436d7f1"} Nov 24 02:32:08 crc kubenswrapper[4888]: I1124 02:32:08.015702 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerStarted","Data":"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817"} Nov 24 02:32:09 crc kubenswrapper[4888]: I1124 02:32:09.029758 4888 generic.go:334] "Generic (PLEG): container finished" podID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerID="69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817" exitCode=0 Nov 24 02:32:09 crc kubenswrapper[4888]: I1124 02:32:09.029845 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerDied","Data":"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817"} Nov 24 02:32:10 crc kubenswrapper[4888]: I1124 02:32:10.040527 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerStarted","Data":"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273"} Nov 24 02:32:10 crc kubenswrapper[4888]: I1124 02:32:10.065729 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2gkxr" podStartSLOduration=2.62062793 podStartE2EDuration="6.065710616s" podCreationTimestamp="2025-11-24 02:32:04 +0000 UTC" firstStartedPulling="2025-11-24 02:32:05.984660397 +0000 UTC m=+7628.567344441" lastFinishedPulling="2025-11-24 02:32:09.429743083 +0000 UTC m=+7632.012427127" observedRunningTime="2025-11-24 02:32:10.064526563 +0000 UTC m=+7632.647210607" watchObservedRunningTime="2025-11-24 02:32:10.065710616 +0000 UTC m=+7632.648394660" Nov 24 02:32:14 crc kubenswrapper[4888]: I1124 02:32:14.876274 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:14 crc kubenswrapper[4888]: I1124 02:32:14.876794 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:15 crc kubenswrapper[4888]: I1124 02:32:15.960345 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-2gkxr" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="registry-server" probeResult="failure" output=< Nov 24 02:32:15 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:32:15 crc kubenswrapper[4888]: > Nov 24 02:32:23 crc kubenswrapper[4888]: I1124 02:32:23.591604 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:32:23 crc kubenswrapper[4888]: I1124 02:32:23.592294 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:32:24 crc kubenswrapper[4888]: I1124 02:32:24.957071 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:25 crc kubenswrapper[4888]: I1124 02:32:25.017713 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:25 crc kubenswrapper[4888]: I1124 02:32:25.201309 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.194433 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2gkxr" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="registry-server" containerID="cri-o://575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273" gracePeriod=2 Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.730314 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.822513 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content\") pod \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.822662 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj8rp\" (UniqueName: \"kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp\") pod \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.822905 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities\") pod \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\" (UID: \"9d264d78-4a5e-42d1-823a-1f13c38db2e2\") " Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.823804 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities" (OuterVolumeSpecName: "utilities") pod "9d264d78-4a5e-42d1-823a-1f13c38db2e2" (UID: "9d264d78-4a5e-42d1-823a-1f13c38db2e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.830545 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp" (OuterVolumeSpecName: "kube-api-access-tj8rp") pod "9d264d78-4a5e-42d1-823a-1f13c38db2e2" (UID: "9d264d78-4a5e-42d1-823a-1f13c38db2e2"). InnerVolumeSpecName "kube-api-access-tj8rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.883287 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d264d78-4a5e-42d1-823a-1f13c38db2e2" (UID: "9d264d78-4a5e-42d1-823a-1f13c38db2e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.925577 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.925608 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d264d78-4a5e-42d1-823a-1f13c38db2e2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:32:26 crc kubenswrapper[4888]: I1124 02:32:26.925620 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj8rp\" (UniqueName: \"kubernetes.io/projected/9d264d78-4a5e-42d1-823a-1f13c38db2e2-kube-api-access-tj8rp\") on node \"crc\" DevicePath \"\"" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.218201 4888 generic.go:334] "Generic (PLEG): container finished" podID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerID="575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273" exitCode=0 Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.218251 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerDied","Data":"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273"} Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.218284 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gkxr" event={"ID":"9d264d78-4a5e-42d1-823a-1f13c38db2e2","Type":"ContainerDied","Data":"08570674a171a77b5b1d6ec50c3794bae2f840d1f17ca1db7995bb609436d7f1"} Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.218306 4888 scope.go:117] "RemoveContainer" containerID="575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.218461 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gkxr" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.260835 4888 scope.go:117] "RemoveContainer" containerID="69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.269969 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.284102 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2gkxr"] Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.285480 4888 scope.go:117] "RemoveContainer" containerID="668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.363069 4888 scope.go:117] "RemoveContainer" containerID="575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273" Nov 24 02:32:27 crc kubenswrapper[4888]: E1124 02:32:27.363563 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273\": container with ID starting with 575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273 not found: ID does not exist" containerID="575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.363592 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273"} err="failed to get container status \"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273\": rpc error: code = NotFound desc = could not find container \"575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273\": container with ID starting with 575d6664c75c3301d6749c9752d2a0924affcd00ca8a8d35f6ee70648ceb9273 not found: ID does not exist" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.363612 4888 scope.go:117] "RemoveContainer" containerID="69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817" Nov 24 02:32:27 crc kubenswrapper[4888]: E1124 02:32:27.363848 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817\": container with ID starting with 69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817 not found: ID does not exist" containerID="69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.363866 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817"} err="failed to get container status \"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817\": rpc error: code = NotFound desc = could not find container \"69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817\": container with ID starting with 69391f340a0ed6d57f8190006348a4c0f345588cfbf55a8d78f3cdf0e2883817 not found: ID does not exist" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.363879 4888 scope.go:117] "RemoveContainer" containerID="668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267" Nov 24 02:32:27 crc kubenswrapper[4888]: E1124 02:32:27.364094 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267\": container with ID starting with 668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267 not found: ID does not exist" containerID="668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267" Nov 24 02:32:27 crc kubenswrapper[4888]: I1124 02:32:27.364112 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267"} err="failed to get container status \"668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267\": rpc error: code = NotFound desc = could not find container \"668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267\": container with ID starting with 668acaf7bc7aecd4a8d4f934f0a90b206eb7f43f5ec7762e663e89fe01a21267 not found: ID does not exist" Nov 24 02:32:28 crc kubenswrapper[4888]: I1124 02:32:28.257175 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" path="/var/lib/kubelet/pods/9d264d78-4a5e-42d1-823a-1f13c38db2e2/volumes" Nov 24 02:32:53 crc kubenswrapper[4888]: I1124 02:32:53.591784 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:32:53 crc kubenswrapper[4888]: I1124 02:32:53.592549 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.040983 4888 scope.go:117] "RemoveContainer" containerID="964a4ec403d0d3242e88140569203f1b5eae7cbb458a12fa63e802a3281a3953" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.592401 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.592582 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.592711 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.594847 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.594953 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" gracePeriod=600 Nov 24 02:33:23 crc kubenswrapper[4888]: E1124 02:33:23.736164 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.921838 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" exitCode=0 Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.921894 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec"} Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.921948 4888 scope.go:117] "RemoveContainer" containerID="ed254bd85cb395f712dde4135fbefcd1c7beca1257c72547221b2b8a2232ab58" Nov 24 02:33:23 crc kubenswrapper[4888]: I1124 02:33:23.922527 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:33:23 crc kubenswrapper[4888]: E1124 02:33:23.922900 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:33:37 crc kubenswrapper[4888]: I1124 02:33:37.245117 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:33:37 crc kubenswrapper[4888]: E1124 02:33:37.245841 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.556021 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:45 crc kubenswrapper[4888]: E1124 02:33:45.557085 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="registry-server" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.557100 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="registry-server" Nov 24 02:33:45 crc kubenswrapper[4888]: E1124 02:33:45.557116 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="extract-content" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.557121 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="extract-content" Nov 24 02:33:45 crc kubenswrapper[4888]: E1124 02:33:45.557132 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="extract-utilities" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.557138 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="extract-utilities" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.557344 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d264d78-4a5e-42d1-823a-1f13c38db2e2" containerName="registry-server" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.561056 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.565808 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.593319 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.593495 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.593539 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjwnx\" (UniqueName: \"kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.694436 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjwnx\" (UniqueName: \"kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.694515 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.694714 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.695252 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.695333 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.716759 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjwnx\" (UniqueName: \"kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx\") pod \"certified-operators-nnnws\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:45 crc kubenswrapper[4888]: I1124 02:33:45.883109 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:46 crc kubenswrapper[4888]: I1124 02:33:46.443222 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:47 crc kubenswrapper[4888]: I1124 02:33:47.223023 4888 generic.go:334] "Generic (PLEG): container finished" podID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerID="03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806" exitCode=0 Nov 24 02:33:47 crc kubenswrapper[4888]: I1124 02:33:47.223226 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerDied","Data":"03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806"} Nov 24 02:33:47 crc kubenswrapper[4888]: I1124 02:33:47.223501 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerStarted","Data":"5b7d9e890e4b97cff26d97b734a6dc3baed4cc0fdbe80e4c002d6fb1a9cfa067"} Nov 24 02:33:48 crc kubenswrapper[4888]: I1124 02:33:48.236794 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerStarted","Data":"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af"} Nov 24 02:33:49 crc kubenswrapper[4888]: I1124 02:33:49.246747 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:33:49 crc kubenswrapper[4888]: E1124 02:33:49.247051 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:33:50 crc kubenswrapper[4888]: I1124 02:33:50.257370 4888 generic.go:334] "Generic (PLEG): container finished" podID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerID="0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af" exitCode=0 Nov 24 02:33:50 crc kubenswrapper[4888]: I1124 02:33:50.260760 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerDied","Data":"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af"} Nov 24 02:33:51 crc kubenswrapper[4888]: I1124 02:33:51.274586 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerStarted","Data":"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f"} Nov 24 02:33:51 crc kubenswrapper[4888]: I1124 02:33:51.302311 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nnnws" podStartSLOduration=2.560003928 podStartE2EDuration="6.302291667s" podCreationTimestamp="2025-11-24 02:33:45 +0000 UTC" firstStartedPulling="2025-11-24 02:33:47.225469507 +0000 UTC m=+7729.808153591" lastFinishedPulling="2025-11-24 02:33:50.967757256 +0000 UTC m=+7733.550441330" observedRunningTime="2025-11-24 02:33:51.294508689 +0000 UTC m=+7733.877192733" watchObservedRunningTime="2025-11-24 02:33:51.302291667 +0000 UTC m=+7733.884975711" Nov 24 02:33:55 crc kubenswrapper[4888]: I1124 02:33:55.883890 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:55 crc kubenswrapper[4888]: I1124 02:33:55.884648 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:55 crc kubenswrapper[4888]: I1124 02:33:55.973031 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.339587 4888 generic.go:334] "Generic (PLEG): container finished" podID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerID="2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0" exitCode=0 Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.339731 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-467dh/must-gather-jk6g7" event={"ID":"ad08b613-2cfe-4f6f-8793-18c3be4885c0","Type":"ContainerDied","Data":"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0"} Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.341047 4888 scope.go:117] "RemoveContainer" containerID="2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0" Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.432984 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.506294 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:56 crc kubenswrapper[4888]: I1124 02:33:56.638785 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-467dh_must-gather-jk6g7_ad08b613-2cfe-4f6f-8793-18c3be4885c0/gather/0.log" Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.358791 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nnnws" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="registry-server" containerID="cri-o://af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f" gracePeriod=2 Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.886101 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.949930 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content\") pod \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.949989 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjwnx\" (UniqueName: \"kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx\") pod \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.950143 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities\") pod \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\" (UID: \"3038ac7f-b07a-4a4e-a748-4ab1846ea039\") " Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.950991 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities" (OuterVolumeSpecName: "utilities") pod "3038ac7f-b07a-4a4e-a748-4ab1846ea039" (UID: "3038ac7f-b07a-4a4e-a748-4ab1846ea039"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.952744 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:33:58 crc kubenswrapper[4888]: I1124 02:33:58.957495 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx" (OuterVolumeSpecName: "kube-api-access-pjwnx") pod "3038ac7f-b07a-4a4e-a748-4ab1846ea039" (UID: "3038ac7f-b07a-4a4e-a748-4ab1846ea039"). InnerVolumeSpecName "kube-api-access-pjwnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.013467 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3038ac7f-b07a-4a4e-a748-4ab1846ea039" (UID: "3038ac7f-b07a-4a4e-a748-4ab1846ea039"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.055731 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3038ac7f-b07a-4a4e-a748-4ab1846ea039-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.055769 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjwnx\" (UniqueName: \"kubernetes.io/projected/3038ac7f-b07a-4a4e-a748-4ab1846ea039-kube-api-access-pjwnx\") on node \"crc\" DevicePath \"\"" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.380465 4888 generic.go:334] "Generic (PLEG): container finished" podID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerID="af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f" exitCode=0 Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.380548 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerDied","Data":"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f"} Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.380585 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnnws" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.380623 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnnws" event={"ID":"3038ac7f-b07a-4a4e-a748-4ab1846ea039","Type":"ContainerDied","Data":"5b7d9e890e4b97cff26d97b734a6dc3baed4cc0fdbe80e4c002d6fb1a9cfa067"} Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.380660 4888 scope.go:117] "RemoveContainer" containerID="af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.420913 4888 scope.go:117] "RemoveContainer" containerID="0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.457071 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.469782 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nnnws"] Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.477773 4888 scope.go:117] "RemoveContainer" containerID="03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.526067 4888 scope.go:117] "RemoveContainer" containerID="af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f" Nov 24 02:33:59 crc kubenswrapper[4888]: E1124 02:33:59.526615 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f\": container with ID starting with af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f not found: ID does not exist" containerID="af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.526643 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f"} err="failed to get container status \"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f\": rpc error: code = NotFound desc = could not find container \"af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f\": container with ID starting with af1c0375e0aa56604b5ed365441cef9470d5cc1e1f1edbf0e8bca43edf8a997f not found: ID does not exist" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.526662 4888 scope.go:117] "RemoveContainer" containerID="0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af" Nov 24 02:33:59 crc kubenswrapper[4888]: E1124 02:33:59.526985 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af\": container with ID starting with 0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af not found: ID does not exist" containerID="0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.527009 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af"} err="failed to get container status \"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af\": rpc error: code = NotFound desc = could not find container \"0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af\": container with ID starting with 0103f1ea3afd0171f25e3525d45cf7b84c4b768667e79aac371c20db489f92af not found: ID does not exist" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.527023 4888 scope.go:117] "RemoveContainer" containerID="03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806" Nov 24 02:33:59 crc kubenswrapper[4888]: E1124 02:33:59.527231 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806\": container with ID starting with 03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806 not found: ID does not exist" containerID="03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806" Nov 24 02:33:59 crc kubenswrapper[4888]: I1124 02:33:59.527252 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806"} err="failed to get container status \"03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806\": rpc error: code = NotFound desc = could not find container \"03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806\": container with ID starting with 03a30f2a314c5a2e442d504ea6b0743e59c4b298d4c48b9b5c0534711c3b8806 not found: ID does not exist" Nov 24 02:34:00 crc kubenswrapper[4888]: I1124 02:34:00.262034 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" path="/var/lib/kubelet/pods/3038ac7f-b07a-4a4e-a748-4ab1846ea039/volumes" Nov 24 02:34:02 crc kubenswrapper[4888]: I1124 02:34:02.245509 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:34:02 crc kubenswrapper[4888]: E1124 02:34:02.246195 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:34:05 crc kubenswrapper[4888]: I1124 02:34:05.431487 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-467dh/must-gather-jk6g7"] Nov 24 02:34:05 crc kubenswrapper[4888]: I1124 02:34:05.432272 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-467dh/must-gather-jk6g7" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="copy" containerID="cri-o://aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48" gracePeriod=2 Nov 24 02:34:05 crc kubenswrapper[4888]: I1124 02:34:05.446290 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-467dh/must-gather-jk6g7"] Nov 24 02:34:05 crc kubenswrapper[4888]: I1124 02:34:05.975248 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-467dh_must-gather-jk6g7_ad08b613-2cfe-4f6f-8793-18c3be4885c0/copy/0.log" Nov 24 02:34:05 crc kubenswrapper[4888]: I1124 02:34:05.976273 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.037680 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output\") pod \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.038119 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kmqc\" (UniqueName: \"kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc\") pod \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\" (UID: \"ad08b613-2cfe-4f6f-8793-18c3be4885c0\") " Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.069488 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc" (OuterVolumeSpecName: "kube-api-access-9kmqc") pod "ad08b613-2cfe-4f6f-8793-18c3be4885c0" (UID: "ad08b613-2cfe-4f6f-8793-18c3be4885c0"). InnerVolumeSpecName "kube-api-access-9kmqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.140622 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kmqc\" (UniqueName: \"kubernetes.io/projected/ad08b613-2cfe-4f6f-8793-18c3be4885c0-kube-api-access-9kmqc\") on node \"crc\" DevicePath \"\"" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.379975 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ad08b613-2cfe-4f6f-8793-18c3be4885c0" (UID: "ad08b613-2cfe-4f6f-8793-18c3be4885c0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.461545 4888 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad08b613-2cfe-4f6f-8793-18c3be4885c0-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.469151 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-467dh_must-gather-jk6g7_ad08b613-2cfe-4f6f-8793-18c3be4885c0/copy/0.log" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.469803 4888 generic.go:334] "Generic (PLEG): container finished" podID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerID="aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48" exitCode=143 Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.469867 4888 scope.go:117] "RemoveContainer" containerID="aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.469904 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-467dh/must-gather-jk6g7" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.494042 4888 scope.go:117] "RemoveContainer" containerID="2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.567133 4888 scope.go:117] "RemoveContainer" containerID="aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48" Nov 24 02:34:06 crc kubenswrapper[4888]: E1124 02:34:06.567638 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48\": container with ID starting with aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48 not found: ID does not exist" containerID="aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.567747 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48"} err="failed to get container status \"aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48\": rpc error: code = NotFound desc = could not find container \"aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48\": container with ID starting with aa25132c671975488cbea774df1a89214fe89d9068caef9183d91af0bc85ba48 not found: ID does not exist" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.567792 4888 scope.go:117] "RemoveContainer" containerID="2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0" Nov 24 02:34:06 crc kubenswrapper[4888]: E1124 02:34:06.568138 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0\": container with ID starting with 2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0 not found: ID does not exist" containerID="2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0" Nov 24 02:34:06 crc kubenswrapper[4888]: I1124 02:34:06.568165 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0"} err="failed to get container status \"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0\": rpc error: code = NotFound desc = could not find container \"2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0\": container with ID starting with 2b3ba8e4a3a5a044f5a7144e5ddd41dccbb0786b27bf41e0229a1e41827daff0 not found: ID does not exist" Nov 24 02:34:08 crc kubenswrapper[4888]: I1124 02:34:08.261406 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" path="/var/lib/kubelet/pods/ad08b613-2cfe-4f6f-8793-18c3be4885c0/volumes" Nov 24 02:34:13 crc kubenswrapper[4888]: I1124 02:34:13.245898 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:34:13 crc kubenswrapper[4888]: E1124 02:34:13.247280 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:34:23 crc kubenswrapper[4888]: I1124 02:34:23.166277 4888 scope.go:117] "RemoveContainer" containerID="fbb272b1ab00299ae0dd66ee785314c74b368576c58d530c568119a702a9ec72" Nov 24 02:34:26 crc kubenswrapper[4888]: I1124 02:34:26.245877 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:34:26 crc kubenswrapper[4888]: E1124 02:34:26.246647 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:34:41 crc kubenswrapper[4888]: I1124 02:34:41.245295 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:34:41 crc kubenswrapper[4888]: E1124 02:34:41.246387 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:34:52 crc kubenswrapper[4888]: I1124 02:34:52.245531 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:34:52 crc kubenswrapper[4888]: E1124 02:34:52.246879 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:35:04 crc kubenswrapper[4888]: I1124 02:35:04.245639 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:35:04 crc kubenswrapper[4888]: E1124 02:35:04.246539 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:35:20 crc kubenswrapper[4888]: I1124 02:35:20.247358 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:35:20 crc kubenswrapper[4888]: E1124 02:35:20.248562 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:35:33 crc kubenswrapper[4888]: I1124 02:35:33.245865 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:35:33 crc kubenswrapper[4888]: E1124 02:35:33.246526 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:35:47 crc kubenswrapper[4888]: I1124 02:35:47.245485 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:35:47 crc kubenswrapper[4888]: E1124 02:35:47.246977 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:36:00 crc kubenswrapper[4888]: I1124 02:36:00.245717 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:36:00 crc kubenswrapper[4888]: E1124 02:36:00.246683 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:36:14 crc kubenswrapper[4888]: I1124 02:36:14.246555 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:36:14 crc kubenswrapper[4888]: E1124 02:36:14.248514 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.525888 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:16 crc kubenswrapper[4888]: E1124 02:36:16.530600 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="copy" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.530802 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="copy" Nov 24 02:36:16 crc kubenswrapper[4888]: E1124 02:36:16.530988 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="extract-content" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.531108 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="extract-content" Nov 24 02:36:16 crc kubenswrapper[4888]: E1124 02:36:16.531245 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="extract-utilities" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.531357 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="extract-utilities" Nov 24 02:36:16 crc kubenswrapper[4888]: E1124 02:36:16.531515 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="registry-server" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.531628 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="registry-server" Nov 24 02:36:16 crc kubenswrapper[4888]: E1124 02:36:16.531751 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="gather" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.531900 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="gather" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.532506 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="gather" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.532687 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad08b613-2cfe-4f6f-8793-18c3be4885c0" containerName="copy" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.532836 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="3038ac7f-b07a-4a4e-a748-4ab1846ea039" containerName="registry-server" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.535463 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.560585 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.716536 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.716609 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.716798 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkpbz\" (UniqueName: \"kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.819253 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.819359 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.819527 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkpbz\" (UniqueName: \"kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.820338 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.820974 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.852107 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkpbz\" (UniqueName: \"kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz\") pod \"redhat-marketplace-rjdml\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:16 crc kubenswrapper[4888]: I1124 02:36:16.858666 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:17 crc kubenswrapper[4888]: I1124 02:36:17.424460 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:18 crc kubenswrapper[4888]: I1124 02:36:18.101159 4888 generic.go:334] "Generic (PLEG): container finished" podID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerID="58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e" exitCode=0 Nov 24 02:36:18 crc kubenswrapper[4888]: I1124 02:36:18.101257 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerDied","Data":"58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e"} Nov 24 02:36:18 crc kubenswrapper[4888]: I1124 02:36:18.101629 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerStarted","Data":"3acb8f618ae5223c09a77fdd8677b0a6ed9e2efdb3d3527b0bd3ba3770f444cf"} Nov 24 02:36:18 crc kubenswrapper[4888]: I1124 02:36:18.104006 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:36:20 crc kubenswrapper[4888]: I1124 02:36:20.130363 4888 generic.go:334] "Generic (PLEG): container finished" podID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerID="9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9" exitCode=0 Nov 24 02:36:20 crc kubenswrapper[4888]: I1124 02:36:20.130427 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerDied","Data":"9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9"} Nov 24 02:36:21 crc kubenswrapper[4888]: I1124 02:36:21.152658 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerStarted","Data":"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e"} Nov 24 02:36:21 crc kubenswrapper[4888]: I1124 02:36:21.193735 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rjdml" podStartSLOduration=2.733917267 podStartE2EDuration="5.193708752s" podCreationTimestamp="2025-11-24 02:36:16 +0000 UTC" firstStartedPulling="2025-11-24 02:36:18.10363606 +0000 UTC m=+7880.686320144" lastFinishedPulling="2025-11-24 02:36:20.563427575 +0000 UTC m=+7883.146111629" observedRunningTime="2025-11-24 02:36:21.17576466 +0000 UTC m=+7883.758448744" watchObservedRunningTime="2025-11-24 02:36:21.193708752 +0000 UTC m=+7883.776392836" Nov 24 02:36:26 crc kubenswrapper[4888]: I1124 02:36:26.858946 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:26 crc kubenswrapper[4888]: I1124 02:36:26.859736 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:26 crc kubenswrapper[4888]: I1124 02:36:26.925374 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:27 crc kubenswrapper[4888]: I1124 02:36:27.276682 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:27 crc kubenswrapper[4888]: I1124 02:36:27.329560 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:28 crc kubenswrapper[4888]: I1124 02:36:28.254117 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:36:28 crc kubenswrapper[4888]: E1124 02:36:28.254668 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.243848 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rjdml" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="registry-server" containerID="cri-o://0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e" gracePeriod=2 Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.787078 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.893272 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content\") pod \"8e13a333-5bf6-45b6-8129-b320cf49d8be\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.893333 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities\") pod \"8e13a333-5bf6-45b6-8129-b320cf49d8be\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.893359 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkpbz\" (UniqueName: \"kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz\") pod \"8e13a333-5bf6-45b6-8129-b320cf49d8be\" (UID: \"8e13a333-5bf6-45b6-8129-b320cf49d8be\") " Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.894561 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities" (OuterVolumeSpecName: "utilities") pod "8e13a333-5bf6-45b6-8129-b320cf49d8be" (UID: "8e13a333-5bf6-45b6-8129-b320cf49d8be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.899147 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz" (OuterVolumeSpecName: "kube-api-access-wkpbz") pod "8e13a333-5bf6-45b6-8129-b320cf49d8be" (UID: "8e13a333-5bf6-45b6-8129-b320cf49d8be"). InnerVolumeSpecName "kube-api-access-wkpbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.914798 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e13a333-5bf6-45b6-8129-b320cf49d8be" (UID: "8e13a333-5bf6-45b6-8129-b320cf49d8be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.995409 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.995437 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e13a333-5bf6-45b6-8129-b320cf49d8be-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:36:29 crc kubenswrapper[4888]: I1124 02:36:29.995447 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkpbz\" (UniqueName: \"kubernetes.io/projected/8e13a333-5bf6-45b6-8129-b320cf49d8be-kube-api-access-wkpbz\") on node \"crc\" DevicePath \"\"" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.268674 4888 generic.go:334] "Generic (PLEG): container finished" podID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerID="0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e" exitCode=0 Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.268789 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjdml" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.275669 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerDied","Data":"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e"} Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.275712 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjdml" event={"ID":"8e13a333-5bf6-45b6-8129-b320cf49d8be","Type":"ContainerDied","Data":"3acb8f618ae5223c09a77fdd8677b0a6ed9e2efdb3d3527b0bd3ba3770f444cf"} Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.275735 4888 scope.go:117] "RemoveContainer" containerID="0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.314120 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.321954 4888 scope.go:117] "RemoveContainer" containerID="9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.343051 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjdml"] Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.348224 4888 scope.go:117] "RemoveContainer" containerID="58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.429099 4888 scope.go:117] "RemoveContainer" containerID="0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e" Nov 24 02:36:30 crc kubenswrapper[4888]: E1124 02:36:30.429463 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e\": container with ID starting with 0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e not found: ID does not exist" containerID="0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.429492 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e"} err="failed to get container status \"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e\": rpc error: code = NotFound desc = could not find container \"0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e\": container with ID starting with 0df7c4d030591d8ce289f50fa4bc886078ab99aaca155ed04bdb4282899fc56e not found: ID does not exist" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.429511 4888 scope.go:117] "RemoveContainer" containerID="9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9" Nov 24 02:36:30 crc kubenswrapper[4888]: E1124 02:36:30.429728 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9\": container with ID starting with 9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9 not found: ID does not exist" containerID="9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.429747 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9"} err="failed to get container status \"9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9\": rpc error: code = NotFound desc = could not find container \"9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9\": container with ID starting with 9919b5ab5f0dd8ea1e95db7947216b1558b71045e5fb3208fd2a80ee38ad6da9 not found: ID does not exist" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.430700 4888 scope.go:117] "RemoveContainer" containerID="58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e" Nov 24 02:36:30 crc kubenswrapper[4888]: E1124 02:36:30.431067 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e\": container with ID starting with 58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e not found: ID does not exist" containerID="58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e" Nov 24 02:36:30 crc kubenswrapper[4888]: I1124 02:36:30.431152 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e"} err="failed to get container status \"58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e\": rpc error: code = NotFound desc = could not find container \"58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e\": container with ID starting with 58d43a6494a6386997446a7874a8e046d5b2f6fa7237ca383718f472fdab7d2e not found: ID does not exist" Nov 24 02:36:32 crc kubenswrapper[4888]: I1124 02:36:32.271409 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" path="/var/lib/kubelet/pods/8e13a333-5bf6-45b6-8129-b320cf49d8be/volumes" Nov 24 02:36:43 crc kubenswrapper[4888]: I1124 02:36:43.247023 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:36:43 crc kubenswrapper[4888]: E1124 02:36:43.249027 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:36:57 crc kubenswrapper[4888]: I1124 02:36:57.246369 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:36:57 crc kubenswrapper[4888]: E1124 02:36:57.247064 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.656540 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x9v9r/must-gather-xlrb5"] Nov 24 02:37:08 crc kubenswrapper[4888]: E1124 02:37:08.657646 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="extract-utilities" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.657661 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="extract-utilities" Nov 24 02:37:08 crc kubenswrapper[4888]: E1124 02:37:08.657687 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="registry-server" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.657693 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="registry-server" Nov 24 02:37:08 crc kubenswrapper[4888]: E1124 02:37:08.657716 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="extract-content" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.657722 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="extract-content" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.657941 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e13a333-5bf6-45b6-8129-b320cf49d8be" containerName="registry-server" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.663325 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.667236 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x9v9r"/"kube-root-ca.crt" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.667564 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x9v9r"/"openshift-service-ca.crt" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.697754 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x9v9r/must-gather-xlrb5"] Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.829845 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzgjh\" (UniqueName: \"kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.830066 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.932654 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.933001 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzgjh\" (UniqueName: \"kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.933292 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.950499 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzgjh\" (UniqueName: \"kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh\") pod \"must-gather-xlrb5\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:08 crc kubenswrapper[4888]: I1124 02:37:08.990865 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:37:09 crc kubenswrapper[4888]: I1124 02:37:09.520193 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x9v9r/must-gather-xlrb5"] Nov 24 02:37:09 crc kubenswrapper[4888]: I1124 02:37:09.844838 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" event={"ID":"294d2afc-c188-4fc5-8f79-f14a0aecefa9","Type":"ContainerStarted","Data":"9f10712cb154f0ed26690353d6e3c89fd65d8fc90281d740bb5cf482b678eca0"} Nov 24 02:37:10 crc kubenswrapper[4888]: I1124 02:37:10.245176 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:37:10 crc kubenswrapper[4888]: E1124 02:37:10.245635 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:37:10 crc kubenswrapper[4888]: I1124 02:37:10.856524 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" event={"ID":"294d2afc-c188-4fc5-8f79-f14a0aecefa9","Type":"ContainerStarted","Data":"eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd"} Nov 24 02:37:10 crc kubenswrapper[4888]: I1124 02:37:10.856588 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" event={"ID":"294d2afc-c188-4fc5-8f79-f14a0aecefa9","Type":"ContainerStarted","Data":"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3"} Nov 24 02:37:10 crc kubenswrapper[4888]: I1124 02:37:10.893845 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" podStartSLOduration=2.893821887 podStartE2EDuration="2.893821887s" podCreationTimestamp="2025-11-24 02:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:37:10.889210345 +0000 UTC m=+7933.471894399" watchObservedRunningTime="2025-11-24 02:37:10.893821887 +0000 UTC m=+7933.476505921" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.457313 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-2jx75"] Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.459873 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.461887 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x9v9r"/"default-dockercfg-t8q2q" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.571742 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.571849 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcxtf\" (UniqueName: \"kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.674137 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.674228 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcxtf\" (UniqueName: \"kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.674312 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.701047 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcxtf\" (UniqueName: \"kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf\") pod \"crc-debug-2jx75\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.784396 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:37:14 crc kubenswrapper[4888]: W1124 02:37:14.832400 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b294a10_9bd3_4fd1_b6ae_b56d79d922db.slice/crio-16cfb0d5b853ab936c3ff8b41c3f75601f49dfdae5b5187102ce052465bb6d64 WatchSource:0}: Error finding container 16cfb0d5b853ab936c3ff8b41c3f75601f49dfdae5b5187102ce052465bb6d64: Status 404 returned error can't find the container with id 16cfb0d5b853ab936c3ff8b41c3f75601f49dfdae5b5187102ce052465bb6d64 Nov 24 02:37:14 crc kubenswrapper[4888]: I1124 02:37:14.912665 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" event={"ID":"7b294a10-9bd3-4fd1-b6ae-b56d79d922db","Type":"ContainerStarted","Data":"16cfb0d5b853ab936c3ff8b41c3f75601f49dfdae5b5187102ce052465bb6d64"} Nov 24 02:37:15 crc kubenswrapper[4888]: I1124 02:37:15.923218 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" event={"ID":"7b294a10-9bd3-4fd1-b6ae-b56d79d922db","Type":"ContainerStarted","Data":"ae391068cb5f2fa738454c43623e3c843baddf40a2ff773ee6d9b88c9fead4cb"} Nov 24 02:37:15 crc kubenswrapper[4888]: I1124 02:37:15.938367 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" podStartSLOduration=1.938346943 podStartE2EDuration="1.938346943s" podCreationTimestamp="2025-11-24 02:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:37:15.932219678 +0000 UTC m=+7938.514903722" watchObservedRunningTime="2025-11-24 02:37:15.938346943 +0000 UTC m=+7938.521030987" Nov 24 02:37:21 crc kubenswrapper[4888]: I1124 02:37:21.244993 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:37:21 crc kubenswrapper[4888]: E1124 02:37:21.245976 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:37:33 crc kubenswrapper[4888]: I1124 02:37:33.246295 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:37:33 crc kubenswrapper[4888]: E1124 02:37:33.246963 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:37:48 crc kubenswrapper[4888]: I1124 02:37:48.252756 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:37:48 crc kubenswrapper[4888]: E1124 02:37:48.253498 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:37:59 crc kubenswrapper[4888]: I1124 02:37:59.332770 4888 generic.go:334] "Generic (PLEG): container finished" podID="7b294a10-9bd3-4fd1-b6ae-b56d79d922db" containerID="ae391068cb5f2fa738454c43623e3c843baddf40a2ff773ee6d9b88c9fead4cb" exitCode=0 Nov 24 02:37:59 crc kubenswrapper[4888]: I1124 02:37:59.333281 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" event={"ID":"7b294a10-9bd3-4fd1-b6ae-b56d79d922db","Type":"ContainerDied","Data":"ae391068cb5f2fa738454c43623e3c843baddf40a2ff773ee6d9b88c9fead4cb"} Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.484688 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.522502 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-2jx75"] Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.533094 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-2jx75"] Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.680150 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host\") pod \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.680283 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host" (OuterVolumeSpecName: "host") pod "7b294a10-9bd3-4fd1-b6ae-b56d79d922db" (UID: "7b294a10-9bd3-4fd1-b6ae-b56d79d922db"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.680338 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcxtf\" (UniqueName: \"kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf\") pod \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\" (UID: \"7b294a10-9bd3-4fd1-b6ae-b56d79d922db\") " Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.681039 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.689190 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf" (OuterVolumeSpecName: "kube-api-access-fcxtf") pod "7b294a10-9bd3-4fd1-b6ae-b56d79d922db" (UID: "7b294a10-9bd3-4fd1-b6ae-b56d79d922db"). InnerVolumeSpecName "kube-api-access-fcxtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:38:00 crc kubenswrapper[4888]: I1124 02:38:00.783916 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcxtf\" (UniqueName: \"kubernetes.io/projected/7b294a10-9bd3-4fd1-b6ae-b56d79d922db-kube-api-access-fcxtf\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.356383 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16cfb0d5b853ab936c3ff8b41c3f75601f49dfdae5b5187102ce052465bb6d64" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.356433 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-2jx75" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.687517 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-lvghr"] Nov 24 02:38:01 crc kubenswrapper[4888]: E1124 02:38:01.688965 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b294a10-9bd3-4fd1-b6ae-b56d79d922db" containerName="container-00" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.689081 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b294a10-9bd3-4fd1-b6ae-b56d79d922db" containerName="container-00" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.689346 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b294a10-9bd3-4fd1-b6ae-b56d79d922db" containerName="container-00" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.690159 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.692002 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x9v9r"/"default-dockercfg-t8q2q" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.704777 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rmdh\" (UniqueName: \"kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.705164 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.806991 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.807308 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rmdh\" (UniqueName: \"kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.807597 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:01 crc kubenswrapper[4888]: I1124 02:38:01.825459 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rmdh\" (UniqueName: \"kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh\") pod \"crc-debug-lvghr\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:02 crc kubenswrapper[4888]: I1124 02:38:02.008173 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:02 crc kubenswrapper[4888]: I1124 02:38:02.270382 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b294a10-9bd3-4fd1-b6ae-b56d79d922db" path="/var/lib/kubelet/pods/7b294a10-9bd3-4fd1-b6ae-b56d79d922db/volumes" Nov 24 02:38:02 crc kubenswrapper[4888]: I1124 02:38:02.367470 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" event={"ID":"d3e77951-74af-450f-a06d-1eba6f739f91","Type":"ContainerStarted","Data":"64f99e22568a7d0ac64d91e232a5d04366542121657bc9ec6eccff010eb91f12"} Nov 24 02:38:02 crc kubenswrapper[4888]: I1124 02:38:02.367523 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" event={"ID":"d3e77951-74af-450f-a06d-1eba6f739f91","Type":"ContainerStarted","Data":"02e474fd41d08a4ac2443382dbb87808b3aee14c001b5396cd64a21d14dc48fa"} Nov 24 02:38:02 crc kubenswrapper[4888]: I1124 02:38:02.386518 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" podStartSLOduration=1.386498618 podStartE2EDuration="1.386498618s" podCreationTimestamp="2025-11-24 02:38:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:38:02.379362164 +0000 UTC m=+7984.962046218" watchObservedRunningTime="2025-11-24 02:38:02.386498618 +0000 UTC m=+7984.969182662" Nov 24 02:38:03 crc kubenswrapper[4888]: I1124 02:38:03.245676 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:38:03 crc kubenswrapper[4888]: E1124 02:38:03.246625 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:38:03 crc kubenswrapper[4888]: I1124 02:38:03.382902 4888 generic.go:334] "Generic (PLEG): container finished" podID="d3e77951-74af-450f-a06d-1eba6f739f91" containerID="64f99e22568a7d0ac64d91e232a5d04366542121657bc9ec6eccff010eb91f12" exitCode=0 Nov 24 02:38:03 crc kubenswrapper[4888]: I1124 02:38:03.382950 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" event={"ID":"d3e77951-74af-450f-a06d-1eba6f739f91","Type":"ContainerDied","Data":"64f99e22568a7d0ac64d91e232a5d04366542121657bc9ec6eccff010eb91f12"} Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.501946 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.665241 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rmdh\" (UniqueName: \"kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh\") pod \"d3e77951-74af-450f-a06d-1eba6f739f91\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.665598 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host\") pod \"d3e77951-74af-450f-a06d-1eba6f739f91\" (UID: \"d3e77951-74af-450f-a06d-1eba6f739f91\") " Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.666384 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host" (OuterVolumeSpecName: "host") pod "d3e77951-74af-450f-a06d-1eba6f739f91" (UID: "d3e77951-74af-450f-a06d-1eba6f739f91"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.672492 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh" (OuterVolumeSpecName: "kube-api-access-4rmdh") pod "d3e77951-74af-450f-a06d-1eba6f739f91" (UID: "d3e77951-74af-450f-a06d-1eba6f739f91"). InnerVolumeSpecName "kube-api-access-4rmdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.767972 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rmdh\" (UniqueName: \"kubernetes.io/projected/d3e77951-74af-450f-a06d-1eba6f739f91-kube-api-access-4rmdh\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:04 crc kubenswrapper[4888]: I1124 02:38:04.768013 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3e77951-74af-450f-a06d-1eba6f739f91-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:05 crc kubenswrapper[4888]: I1124 02:38:05.081014 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-lvghr"] Nov 24 02:38:05 crc kubenswrapper[4888]: I1124 02:38:05.090860 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-lvghr"] Nov 24 02:38:05 crc kubenswrapper[4888]: I1124 02:38:05.413985 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e474fd41d08a4ac2443382dbb87808b3aee14c001b5396cd64a21d14dc48fa" Nov 24 02:38:05 crc kubenswrapper[4888]: I1124 02:38:05.414315 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-lvghr" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.273460 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e77951-74af-450f-a06d-1eba6f739f91" path="/var/lib/kubelet/pods/d3e77951-74af-450f-a06d-1eba6f739f91/volumes" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.298315 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-7b59x"] Nov 24 02:38:06 crc kubenswrapper[4888]: E1124 02:38:06.298781 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e77951-74af-450f-a06d-1eba6f739f91" containerName="container-00" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.298795 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e77951-74af-450f-a06d-1eba6f739f91" containerName="container-00" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.298998 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e77951-74af-450f-a06d-1eba6f739f91" containerName="container-00" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.299718 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.302410 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x9v9r"/"default-dockercfg-t8q2q" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.305426 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.305944 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkbpq\" (UniqueName: \"kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.407161 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkbpq\" (UniqueName: \"kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.407291 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.407382 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.437925 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkbpq\" (UniqueName: \"kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq\") pod \"crc-debug-7b59x\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: I1124 02:38:06.624288 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:06 crc kubenswrapper[4888]: W1124 02:38:06.654484 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5684cc1d_440c_46e6_9cbc_559322e12404.slice/crio-b933d40d13f3f56afa58860180424bb83e4bff1ecd6e9d06bc5afe28ed4e1e83 WatchSource:0}: Error finding container b933d40d13f3f56afa58860180424bb83e4bff1ecd6e9d06bc5afe28ed4e1e83: Status 404 returned error can't find the container with id b933d40d13f3f56afa58860180424bb83e4bff1ecd6e9d06bc5afe28ed4e1e83 Nov 24 02:38:07 crc kubenswrapper[4888]: I1124 02:38:07.435155 4888 generic.go:334] "Generic (PLEG): container finished" podID="5684cc1d-440c-46e6-9cbc-559322e12404" containerID="7934902b4d638c675e6408955bef2a67fffac25804ef606f3a1ddbe741c2a15b" exitCode=0 Nov 24 02:38:07 crc kubenswrapper[4888]: I1124 02:38:07.435235 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" event={"ID":"5684cc1d-440c-46e6-9cbc-559322e12404","Type":"ContainerDied","Data":"7934902b4d638c675e6408955bef2a67fffac25804ef606f3a1ddbe741c2a15b"} Nov 24 02:38:07 crc kubenswrapper[4888]: I1124 02:38:07.435584 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" event={"ID":"5684cc1d-440c-46e6-9cbc-559322e12404","Type":"ContainerStarted","Data":"b933d40d13f3f56afa58860180424bb83e4bff1ecd6e9d06bc5afe28ed4e1e83"} Nov 24 02:38:07 crc kubenswrapper[4888]: I1124 02:38:07.480439 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-7b59x"] Nov 24 02:38:07 crc kubenswrapper[4888]: I1124 02:38:07.491083 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x9v9r/crc-debug-7b59x"] Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.557691 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.756302 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host\") pod \"5684cc1d-440c-46e6-9cbc-559322e12404\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.756409 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkbpq\" (UniqueName: \"kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq\") pod \"5684cc1d-440c-46e6-9cbc-559322e12404\" (UID: \"5684cc1d-440c-46e6-9cbc-559322e12404\") " Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.756501 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host" (OuterVolumeSpecName: "host") pod "5684cc1d-440c-46e6-9cbc-559322e12404" (UID: "5684cc1d-440c-46e6-9cbc-559322e12404"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.757996 4888 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5684cc1d-440c-46e6-9cbc-559322e12404-host\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.763571 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq" (OuterVolumeSpecName: "kube-api-access-kkbpq") pod "5684cc1d-440c-46e6-9cbc-559322e12404" (UID: "5684cc1d-440c-46e6-9cbc-559322e12404"). InnerVolumeSpecName "kube-api-access-kkbpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:38:08 crc kubenswrapper[4888]: I1124 02:38:08.861528 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkbpq\" (UniqueName: \"kubernetes.io/projected/5684cc1d-440c-46e6-9cbc-559322e12404-kube-api-access-kkbpq\") on node \"crc\" DevicePath \"\"" Nov 24 02:38:09 crc kubenswrapper[4888]: I1124 02:38:09.459231 4888 scope.go:117] "RemoveContainer" containerID="7934902b4d638c675e6408955bef2a67fffac25804ef606f3a1ddbe741c2a15b" Nov 24 02:38:09 crc kubenswrapper[4888]: I1124 02:38:09.459285 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/crc-debug-7b59x" Nov 24 02:38:10 crc kubenswrapper[4888]: I1124 02:38:10.268923 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5684cc1d-440c-46e6-9cbc-559322e12404" path="/var/lib/kubelet/pods/5684cc1d-440c-46e6-9cbc-559322e12404/volumes" Nov 24 02:38:17 crc kubenswrapper[4888]: I1124 02:38:17.245237 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:38:17 crc kubenswrapper[4888]: E1124 02:38:17.246464 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:38:28 crc kubenswrapper[4888]: I1124 02:38:28.262357 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:38:28 crc kubenswrapper[4888]: I1124 02:38:28.706844 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae"} Nov 24 02:39:10 crc kubenswrapper[4888]: I1124 02:39:10.938597 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-api/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.032495 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-evaluator/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.106060 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-listener/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.111140 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_24be4c10-1e0a-4452-b76b-1d979796231d/aodh-notifier/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.216502 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dcbbbf466-gt2vf_8e2c8c4b-730c-4c00-9dcd-03dd6b98b288/barbican-api/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.426014 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dcbbbf466-gt2vf_8e2c8c4b-730c-4c00-9dcd-03dd6b98b288/barbican-api-log/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.529098 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-96d8b7cb6-6rvkq_f68b1099-7b2b-4d27-961c-ac412f438292/barbican-keystone-listener/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.708347 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-847b7f6b9f-ngtnz_b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf/barbican-worker/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.714620 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-96d8b7cb6-6rvkq_f68b1099-7b2b-4d27-961c-ac412f438292/barbican-keystone-listener-log/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.735323 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-847b7f6b9f-ngtnz_b3c800fe-3dd6-47f6-b6b6-0dd825ed86bf/barbican-worker-log/0.log" Nov 24 02:39:11 crc kubenswrapper[4888]: I1124 02:39:11.933102 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-49xnt_ebbc9282-a85b-4484-b130-5b0c91030009/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.030427 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/ceilometer-central-agent/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.111394 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/proxy-httpd/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.159752 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/sg-core/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.165008 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9ea8705f-eabc-47ad-9f38-c77011ba6616/ceilometer-notification-agent/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.309704 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-lnnjv_ea89083c-a0ce-4bfe-b419-be89f9643924/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.405540 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xhdds_99022481-7ff1-4b5a-a06e-376fbeaa29ba/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.623067 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6a2e8920-d187-4deb-85c8-b5b831af5995/cinder-api/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.674550 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6a2e8920-d187-4deb-85c8-b5b831af5995/cinder-api-log/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.866523 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5a08f210-f6e8-46f3-9796-8f4236a53dd1/probe/0.log" Nov 24 02:39:12 crc kubenswrapper[4888]: I1124 02:39:12.981686 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5a08f210-f6e8-46f3-9796-8f4236a53dd1/cinder-backup/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.017066 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_32d98628-e960-4e56-b620-7ba6050faadf/cinder-scheduler/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.113055 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_32d98628-e960-4e56-b620-7ba6050faadf/probe/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.229666 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a2644986-db7d-4368-956b-65dc26585e3b/cinder-volume/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.335850 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a2644986-db7d-4368-956b-65dc26585e3b/probe/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.387272 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-r99nj_5f1ab899-c7e5-4618-b282-8eebc0dec80b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.538779 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lwvbr_32bcbd1c-35f4-4632-ab42-01e6cc578d93/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.632089 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/init/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.804175 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/init/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.861870 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-k75jx_9d8de233-4a00-482c-b35f-170370015b11/dnsmasq-dns/0.log" Nov 24 02:39:13 crc kubenswrapper[4888]: I1124 02:39:13.893367 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5296f4e7-f819-43bb-9566-ee12ba6d51f0/glance-httpd/0.log" Nov 24 02:39:14 crc kubenswrapper[4888]: I1124 02:39:14.040196 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5296f4e7-f819-43bb-9566-ee12ba6d51f0/glance-log/0.log" Nov 24 02:39:14 crc kubenswrapper[4888]: I1124 02:39:14.121585 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2ae1990-d967-4fe3-80cf-9520e5b83286/glance-httpd/0.log" Nov 24 02:39:14 crc kubenswrapper[4888]: I1124 02:39:14.122524 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2ae1990-d967-4fe3-80cf-9520e5b83286/glance-log/0.log" Nov 24 02:39:14 crc kubenswrapper[4888]: I1124 02:39:14.626846 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-799c49dfcf-w6lnt_ef846e73-7a80-4d25-ac41-cfa95d68cc1e/heat-engine/0.log" Nov 24 02:39:14 crc kubenswrapper[4888]: I1124 02:39:14.805155 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7dfb67ff5b-2cglk_78ef5328-d894-4921-8464-62bdaf1c4208/horizon/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.008579 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pk5zv_7fce69b0-15d9-41d7-94f0-2263b74dff10/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.257339 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-rkgqz_e5f0ab5f-207b-4e2b-826a-87daf6aa5f25/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.578550 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-775b95987-88pxg_eb06d977-15f5-4263-aa7e-4fa2b2964847/heat-cfnapi/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.633522 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7dfb67ff5b-2cglk_78ef5328-d894-4921-8464-62bdaf1c4208/horizon-log/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.653168 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7f5f5cb65-zfpq5_6baa14da-0b8b-43da-81b9-605e596cbb11/heat-api/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.777255 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399101-qdc8s_81ab4fd4-2f1c-4023-b7a1-528e59bb56ad/keystone-cron/0.log" Nov 24 02:39:15 crc kubenswrapper[4888]: I1124 02:39:15.937281 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399161-vzqj4_5ab4f95d-24df-4a76-b678-a62601ae4375/keystone-cron/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.037784 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fe0c6981-4d50-4dba-b7a4-cabf7beb1947/kube-state-metrics/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.173200 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b7f898bdd-bbfm7_b5d4d09b-6dbb-4318-9ab2-fbb52b23cf10/keystone-api/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.182437 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-7xwh4_c4e89186-7454-4e0c-85ed-f7c90f6a7b0f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.353760 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-cvzdf_91206204-242c-4c3d-b686-9a4beebb76f1/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.478875 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8bcac458-0aa6-4a6c-8586-666803573def/manila-api-log/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.539392 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8bcac458-0aa6-4a6c-8586-666803573def/manila-api/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.681124 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eff408f2-f760-40a0-9016-b149c398cf85/manila-scheduler/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.693246 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eff408f2-f760-40a0-9016-b149c398cf85/probe/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.836491 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5f42ea1-62ff-4f74-bf41-4425b7332de8/probe/0.log" Nov 24 02:39:16 crc kubenswrapper[4888]: I1124 02:39:16.855453 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5f42ea1-62ff-4f74-bf41-4425b7332de8/manila-share/0.log" Nov 24 02:39:17 crc kubenswrapper[4888]: I1124 02:39:17.048936 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_f25aa6c8-4f87-4d83-8563-fc22a4bfa695/mysqld-exporter/0.log" Nov 24 02:39:17 crc kubenswrapper[4888]: I1124 02:39:17.388079 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zmvv_fb3242a5-393f-4fda-a791-0d1246a5bded/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:17 crc kubenswrapper[4888]: I1124 02:39:17.500147 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-686d6d8cd9-m24dl_d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16/neutron-httpd/0.log" Nov 24 02:39:17 crc kubenswrapper[4888]: I1124 02:39:17.552844 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-686d6d8cd9-m24dl_d7c54d11-cfd0-4ff1-b628-03f5bd6a6a16/neutron-api/0.log" Nov 24 02:39:18 crc kubenswrapper[4888]: I1124 02:39:18.058042 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_be8f5419-cba3-4446-937d-f8d4b06c1c0e/nova-cell0-conductor-conductor/0.log" Nov 24 02:39:18 crc kubenswrapper[4888]: I1124 02:39:18.448327 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f5cc70d4-9318-4552-825b-9ad495d378b0/nova-cell1-conductor-conductor/0.log" Nov 24 02:39:18 crc kubenswrapper[4888]: I1124 02:39:18.458357 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f4060c08-c47b-483a-8a17-661a171187f2/nova-api-log/0.log" Nov 24 02:39:18 crc kubenswrapper[4888]: I1124 02:39:18.763535 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a5e4972d-5b76-4f46-a9fa-3a4425534ad8/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 02:39:18 crc kubenswrapper[4888]: I1124 02:39:18.763834 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-zdbbt_59680387-8e94-4e51-9b44-9c0829e262a6/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.141516 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfb6aad7-f065-42f1-9021-f5debe134c7d/nova-metadata-log/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.224871 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f4060c08-c47b-483a-8a17-661a171187f2/nova-api-api/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.500475 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/mysql-bootstrap/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.607472 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_894a3216-d75f-4229-849f-4b7f4e942d96/nova-scheduler-scheduler/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.687918 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/mysql-bootstrap/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.753027 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a963d44-99cd-45d4-afd8-9c9c7bfb95e7/galera/0.log" Nov 24 02:39:19 crc kubenswrapper[4888]: I1124 02:39:19.901999 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/mysql-bootstrap/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.106779 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/mysql-bootstrap/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.151517 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1c29616-6ecc-4f98-9849-b04242cf0a66/galera/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.292618 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a036e658-146d-4eb0-a197-939e6bbc31f8/openstackclient/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.421597 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4s282_6f28f199-4a0d-4c00-8aef-c5219f6fcc10/openstack-network-exporter/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.632625 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server-init/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.760695 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server-init/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.836450 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovs-vswitchd/0.log" Nov 24 02:39:20 crc kubenswrapper[4888]: I1124 02:39:20.855405 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6h6d_fb427d6c-c0ae-454c-8c53-71d20209ea47/ovsdb-server/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.044288 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x5tvh_053aa355-b781-49f4-8e8a-80ae64c4de62/ovn-controller/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.242128 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gjwnp_26532baf-63c3-4d2f-87c2-9c690c53823c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.320129 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c37246a-89a8-4f28-95f9-149b6c9f1790/openstack-network-exporter/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.503991 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c37246a-89a8-4f28-95f9-149b6c9f1790/ovn-northd/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.580884 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_728a0118-33ee-4fb3-b8e2-198e075adc31/openstack-network-exporter/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.714074 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_728a0118-33ee-4fb3-b8e2-198e075adc31/ovsdbserver-nb/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.800277 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cdce96a2-6d42-4761-9170-ef61211cfc92/openstack-network-exporter/0.log" Nov 24 02:39:21 crc kubenswrapper[4888]: I1124 02:39:21.952803 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cdce96a2-6d42-4761-9170-ef61211cfc92/ovsdbserver-sb/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.108310 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfb6aad7-f065-42f1-9021-f5debe134c7d/nova-metadata-metadata/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.185607 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-669f454766-khh4h_4c22d012-f519-4556-887a-2a0a11d7f3ec/placement-api/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.308391 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/init-config-reloader/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.328509 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-669f454766-khh4h_4c22d012-f519-4556-887a-2a0a11d7f3ec/placement-log/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.510488 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/init-config-reloader/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.519143 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/config-reloader/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.530464 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/thanos-sidecar/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.573795 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7ba9e0fb-846a-49dc-bdc3-a0e201324dab/prometheus/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.732445 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/setup-container/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.912870 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/setup-container/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.939281 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f987cf1a-247e-411e-9d03-4a5f44a98d14/rabbitmq/0.log" Nov 24 02:39:22 crc kubenswrapper[4888]: I1124 02:39:22.957003 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/setup-container/0.log" Nov 24 02:39:23 crc kubenswrapper[4888]: I1124 02:39:23.284714 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/rabbitmq/0.log" Nov 24 02:39:23 crc kubenswrapper[4888]: I1124 02:39:23.328194 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_021b7361-39f2-47fa-b028-b1656b5fe94c/setup-container/0.log" Nov 24 02:39:23 crc kubenswrapper[4888]: I1124 02:39:23.359164 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-w6vll_2d25d7f9-02eb-4f78-a3c9-2f9a9a533797/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:23 crc kubenswrapper[4888]: I1124 02:39:23.739446 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m2twt_30e410dc-697f-4714-9e4b-a2f8a71796c4/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:23 crc kubenswrapper[4888]: I1124 02:39:23.768417 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jrjs2_bab960db-54de-4def-b56e-9f31b8df0e03/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.051901 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pbl78_8804f688-d740-4f33-ba57-ac0ca3d63a2b/ssh-known-hosts-edpm-deployment/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.243379 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6dd96dd9bc-c9x6n_07c3d65f-a35c-481d-a773-a9dc4dd914b5/proxy-server/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.320894 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bq8p7_1eaecd0e-0a6f-45d1-95da-c6ed86192911/swift-ring-rebalance/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.399858 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6dd96dd9bc-c9x6n_07c3d65f-a35c-481d-a773-a9dc4dd914b5/proxy-httpd/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.529317 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-auditor/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.616274 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-reaper/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.709319 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-replicator/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.749754 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/account-server/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.852172 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-auditor/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.860889 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-replicator/0.log" Nov 24 02:39:24 crc kubenswrapper[4888]: I1124 02:39:24.934044 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-server/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.020347 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/container-updater/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.083824 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-auditor/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.131748 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-expirer/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.237321 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-replicator/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.250172 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-server/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.298948 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/object-updater/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.347673 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/rsync/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.436451 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9f931286-c309-4f90-8c3e-7f2b3682e68c/swift-recon-cron/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.558947 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-t54vc_7009799b-7622-43c3-856d-7a368de7608f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.680438 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-jjnc9_1fb69384-3ad1-424b-93ad-2716c502bc4d/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:25 crc kubenswrapper[4888]: I1124 02:39:25.941204 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f05e2e09-1cf7-4c0f-b2d3-7ac374e3f717/test-operator-logs-container/0.log" Nov 24 02:39:26 crc kubenswrapper[4888]: I1124 02:39:26.061210 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-crwxl_de6be747-eec3-445b-a6e5-3cd034b68c6a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 02:39:26 crc kubenswrapper[4888]: I1124 02:39:26.798142 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cc931795-8f17-4afd-a16c-264855b1755c/tempest-tests-tempest-tests-runner/0.log" Nov 24 02:39:42 crc kubenswrapper[4888]: I1124 02:39:42.500855 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fca78ec3-fec9-4c3e-aee9-75d3e049c668/memcached/0.log" Nov 24 02:39:55 crc kubenswrapper[4888]: I1124 02:39:55.995419 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.179950 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.202738 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.216110 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.380204 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/pull/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.399252 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/util/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.457781 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_af23c382836ffb5cca6f9087fa0d3058bc02d07945399484d8411431fet7cxv_de3cc12f-3ea8-4ea0-bbba-694b5a3258b4/extract/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.541476 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-gphtr_e7342ffc-1f02-4eca-9be9-5816f5f0c86c/kube-rbac-proxy/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.633741 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-4gh28_5fe80165-08e7-4a34-8290-fd0d7fd22dd2/kube-rbac-proxy/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.665982 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-gphtr_e7342ffc-1f02-4eca-9be9-5816f5f0c86c/manager/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.777322 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-4gh28_5fe80165-08e7-4a34-8290-fd0d7fd22dd2/manager/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.853340 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-9nb4l_a277e2cc-4990-4d79-8a4f-ec03c81d747d/kube-rbac-proxy/0.log" Nov 24 02:39:56 crc kubenswrapper[4888]: I1124 02:39:56.877691 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-9nb4l_a277e2cc-4990-4d79-8a4f-ec03c81d747d/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.037384 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-mtqr5_a701224c-a2ad-49b6-98b7-55fd37051e5b/kube-rbac-proxy/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.172983 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-mtqr5_a701224c-a2ad-49b6-98b7-55fd37051e5b/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.217738 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-wbxrh_6cf992a1-eac0-4269-b8bb-e14554116727/kube-rbac-proxy/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.343355 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-wbxrh_6cf992a1-eac0-4269-b8bb-e14554116727/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.368670 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-llt2h_10a23051-0913-44fd-abda-89436a7e3b59/kube-rbac-proxy/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.407481 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-llt2h_10a23051-0913-44fd-abda-89436a7e3b59/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.526078 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-zl954_e6420e55-4951-4546-b252-44c25d49a0c3/kube-rbac-proxy/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.654252 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jk6ph_3766dfec-2de1-4df2-8609-0cbbe8599342/kube-rbac-proxy/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.737553 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-zl954_e6420e55-4951-4546-b252-44c25d49a0c3/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.753253 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jk6ph_3766dfec-2de1-4df2-8609-0cbbe8599342/manager/0.log" Nov 24 02:39:57 crc kubenswrapper[4888]: I1124 02:39:57.860469 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-8kqtd_587fee04-3081-4c59-8125-0e47436e2347/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.011013 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-8kqtd_587fee04-3081-4c59-8125-0e47436e2347/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.068794 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-tsb9j_c18797ad-a3dc-46af-9aaf-d9e17918228a/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.107698 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-tsb9j_c18797ad-a3dc-46af-9aaf-d9e17918228a/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.227692 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-4pmdz_dbaba630-daf7-46c3-807a-6e76128fc7d6/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.276393 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-4pmdz_dbaba630-daf7-46c3-807a-6e76128fc7d6/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.416248 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-fn4cs_afa14c5d-d820-4b05-bddd-4d00e49a421a/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.516525 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-fn4cs_afa14c5d-d820-4b05-bddd-4d00e49a421a/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.544267 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-7zjtv_f4bc41bd-b9d5-4acf-978e-75144d8b43a3/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.703300 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-vxffn_750c449f-a775-4ee8-82b8-a6be531250fe/kube-rbac-proxy/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.730459 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-7zjtv_f4bc41bd-b9d5-4acf-978e-75144d8b43a3/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.773838 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-vxffn_750c449f-a775-4ee8-82b8-a6be531250fe/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.924928 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh_2b35fb96-9ea1-4aaa-9396-cd5f705f63bb/manager/0.log" Nov 24 02:39:58 crc kubenswrapper[4888]: I1124 02:39:58.943342 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-wr4hh_2b35fb96-9ea1-4aaa-9396-cd5f705f63bb/kube-rbac-proxy/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.159857 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d84888d56-jk5d2_db75890a-4c13-4f98-a410-aa4f98890ccc/kube-rbac-proxy/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.294184 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75b7b5b9df-7j8cf_9cc98d14-8e44-4556-b00a-ed2069d395cf/kube-rbac-proxy/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.500104 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ggr8j_927e1330-0a2c-4e33-b97d-5de99ce8138c/registry-server/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.553622 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75b7b5b9df-7j8cf_9cc98d14-8e44-4556-b00a-ed2069d395cf/operator/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.756906 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-c9cfn_8a57fb36-c0db-4778-9c17-7241b5e8f602/kube-rbac-proxy/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.768087 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-c9cfn_8a57fb36-c0db-4778-9c17-7241b5e8f602/manager/0.log" Nov 24 02:39:59 crc kubenswrapper[4888]: I1124 02:39:59.851760 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-98b6n_fef147fd-ee85-4c05-9652-5e44223a3e5e/kube-rbac-proxy/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.016751 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-98b6n_fef147fd-ee85-4c05-9652-5e44223a3e5e/manager/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.086357 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-m2vgv_e569ce1f-9272-4c6a-bb91-a01616bb5883/operator/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.226158 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-5ck2h_3a5b66a7-1cf6-4840-870c-f7a6968bd08a/kube-rbac-proxy/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.262721 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-5ck2h_3a5b66a7-1cf6-4840-870c-f7a6968bd08a/manager/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.331206 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67c4db9997-7rqlj_bc56a88d-e719-40f5-a210-1b7afdffae8b/kube-rbac-proxy/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.511221 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-mn678_82f2609b-e676-49f8-9ab4-d06fb3e626be/kube-rbac-proxy/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.596718 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-mn678_82f2609b-e676-49f8-9ab4-d06fb3e626be/manager/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.867103 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-6mmtv_c36e3742-c48a-48d1-96c8-87c3809a4183/manager/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.874429 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-6mmtv_c36e3742-c48a-48d1-96c8-87c3809a4183/kube-rbac-proxy/0.log" Nov 24 02:40:00 crc kubenswrapper[4888]: I1124 02:40:00.929397 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67c4db9997-7rqlj_bc56a88d-e719-40f5-a210-1b7afdffae8b/manager/0.log" Nov 24 02:40:01 crc kubenswrapper[4888]: I1124 02:40:01.028865 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d84888d56-jk5d2_db75890a-4c13-4f98-a410-aa4f98890ccc/manager/0.log" Nov 24 02:40:17 crc kubenswrapper[4888]: I1124 02:40:17.746293 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j6829_814a1313-1c15-436c-9a1d-132859fabdea/control-plane-machine-set-operator/0.log" Nov 24 02:40:17 crc kubenswrapper[4888]: I1124 02:40:17.941319 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8s79_103c3930-7610-4c38-af13-cb0f06da2629/machine-api-operator/0.log" Nov 24 02:40:17 crc kubenswrapper[4888]: I1124 02:40:17.948765 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8s79_103c3930-7610-4c38-af13-cb0f06da2629/kube-rbac-proxy/0.log" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.072368 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:40:30 crc kubenswrapper[4888]: E1124 02:40:30.073562 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5684cc1d-440c-46e6-9cbc-559322e12404" containerName="container-00" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.073581 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="5684cc1d-440c-46e6-9cbc-559322e12404" containerName="container-00" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.073903 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="5684cc1d-440c-46e6-9cbc-559322e12404" containerName="container-00" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.075956 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.099711 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.158258 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.158312 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.158519 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dgx\" (UniqueName: \"kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.260233 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dgx\" (UniqueName: \"kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.260917 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.260958 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.261323 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.261374 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.284940 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dgx\" (UniqueName: \"kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx\") pod \"redhat-operators-94jzf\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.412158 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:30 crc kubenswrapper[4888]: I1124 02:40:30.944225 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:40:31 crc kubenswrapper[4888]: I1124 02:40:31.021289 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerStarted","Data":"0b02f07fe8c647d984ecc1b29e6e24e644df66c39fe73d1fb41c48291189808b"} Nov 24 02:40:31 crc kubenswrapper[4888]: I1124 02:40:31.703317 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2ls5n_087d3537-623e-4f8b-be5b-b9342a2cf19f/cert-manager-controller/0.log" Nov 24 02:40:31 crc kubenswrapper[4888]: I1124 02:40:31.859414 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vjmm8_25c1c650-be40-437b-9ea3-8e9a60b2f2f2/cert-manager-cainjector/0.log" Nov 24 02:40:31 crc kubenswrapper[4888]: I1124 02:40:31.879959 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-g8kdr_6a61128e-aa42-4201-87ae-856cc4e6f682/cert-manager-webhook/0.log" Nov 24 02:40:32 crc kubenswrapper[4888]: I1124 02:40:32.034595 4888 generic.go:334] "Generic (PLEG): container finished" podID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerID="4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff" exitCode=0 Nov 24 02:40:32 crc kubenswrapper[4888]: I1124 02:40:32.034664 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerDied","Data":"4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff"} Nov 24 02:40:33 crc kubenswrapper[4888]: I1124 02:40:33.043921 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerStarted","Data":"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095"} Nov 24 02:40:37 crc kubenswrapper[4888]: I1124 02:40:37.102817 4888 generic.go:334] "Generic (PLEG): container finished" podID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerID="322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095" exitCode=0 Nov 24 02:40:37 crc kubenswrapper[4888]: I1124 02:40:37.102869 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerDied","Data":"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095"} Nov 24 02:40:38 crc kubenswrapper[4888]: I1124 02:40:38.118565 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerStarted","Data":"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d"} Nov 24 02:40:38 crc kubenswrapper[4888]: I1124 02:40:38.139356 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-94jzf" podStartSLOduration=2.289562837 podStartE2EDuration="8.139338392s" podCreationTimestamp="2025-11-24 02:40:30 +0000 UTC" firstStartedPulling="2025-11-24 02:40:32.036286689 +0000 UTC m=+8134.618970743" lastFinishedPulling="2025-11-24 02:40:37.886062254 +0000 UTC m=+8140.468746298" observedRunningTime="2025-11-24 02:40:38.137438978 +0000 UTC m=+8140.720123032" watchObservedRunningTime="2025-11-24 02:40:38.139338392 +0000 UTC m=+8140.722022436" Nov 24 02:40:40 crc kubenswrapper[4888]: I1124 02:40:40.413204 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:40 crc kubenswrapper[4888]: I1124 02:40:40.413605 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:40:41 crc kubenswrapper[4888]: I1124 02:40:41.480445 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-94jzf" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" probeResult="failure" output=< Nov 24 02:40:41 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:40:41 crc kubenswrapper[4888]: > Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.431255 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-vzwm2_d88df599-7094-4325-afeb-d4188a9b948c/nmstate-console-plugin/0.log" Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.608244 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fdwl9_00daa12e-54c4-46e9-835c-0a2d2c9ad90e/nmstate-handler/0.log" Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.651765 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kgwmb_f9bc22c3-35a5-4afc-90fe-67b9c086b46a/kube-rbac-proxy/0.log" Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.654548 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kgwmb_f9bc22c3-35a5-4afc-90fe-67b9c086b46a/nmstate-metrics/0.log" Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.859953 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-mncvc_f1648233-9c9f-4a7a-8acc-35d91c1a0b04/nmstate-operator/0.log" Nov 24 02:40:45 crc kubenswrapper[4888]: I1124 02:40:45.908059 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-fhrr5_2737508c-e47e-4ae4-b412-b665096fbf78/nmstate-webhook/0.log" Nov 24 02:40:51 crc kubenswrapper[4888]: I1124 02:40:51.467371 4888 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-94jzf" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" probeResult="failure" output=< Nov 24 02:40:51 crc kubenswrapper[4888]: timeout: failed to connect service ":50051" within 1s Nov 24 02:40:51 crc kubenswrapper[4888]: > Nov 24 02:40:53 crc kubenswrapper[4888]: I1124 02:40:53.592476 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:40:53 crc kubenswrapper[4888]: I1124 02:40:53.593067 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:40:58 crc kubenswrapper[4888]: I1124 02:40:58.707085 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/kube-rbac-proxy/0.log" Nov 24 02:40:58 crc kubenswrapper[4888]: I1124 02:40:58.732067 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/manager/0.log" Nov 24 02:41:00 crc kubenswrapper[4888]: I1124 02:41:00.505573 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:41:00 crc kubenswrapper[4888]: I1124 02:41:00.587644 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:41:01 crc kubenswrapper[4888]: I1124 02:41:01.273565 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:41:02 crc kubenswrapper[4888]: I1124 02:41:02.391776 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-94jzf" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" containerID="cri-o://6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d" gracePeriod=2 Nov 24 02:41:02 crc kubenswrapper[4888]: I1124 02:41:02.927632 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.076241 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities\") pod \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.076325 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4dgx\" (UniqueName: \"kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx\") pod \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.076473 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content\") pod \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\" (UID: \"dd2a6cd6-8d55-4247-83dd-c5b69975071b\") " Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.077624 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities" (OuterVolumeSpecName: "utilities") pod "dd2a6cd6-8d55-4247-83dd-c5b69975071b" (UID: "dd2a6cd6-8d55-4247-83dd-c5b69975071b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.086642 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx" (OuterVolumeSpecName: "kube-api-access-d4dgx") pod "dd2a6cd6-8d55-4247-83dd-c5b69975071b" (UID: "dd2a6cd6-8d55-4247-83dd-c5b69975071b"). InnerVolumeSpecName "kube-api-access-d4dgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.178895 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4dgx\" (UniqueName: \"kubernetes.io/projected/dd2a6cd6-8d55-4247-83dd-c5b69975071b-kube-api-access-d4dgx\") on node \"crc\" DevicePath \"\"" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.178938 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.180288 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd2a6cd6-8d55-4247-83dd-c5b69975071b" (UID: "dd2a6cd6-8d55-4247-83dd-c5b69975071b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.282174 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2a6cd6-8d55-4247-83dd-c5b69975071b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.405086 4888 generic.go:334] "Generic (PLEG): container finished" podID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerID="6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d" exitCode=0 Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.405132 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerDied","Data":"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d"} Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.405139 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94jzf" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.405162 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94jzf" event={"ID":"dd2a6cd6-8d55-4247-83dd-c5b69975071b","Type":"ContainerDied","Data":"0b02f07fe8c647d984ecc1b29e6e24e644df66c39fe73d1fb41c48291189808b"} Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.405183 4888 scope.go:117] "RemoveContainer" containerID="6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.442514 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.443792 4888 scope.go:117] "RemoveContainer" containerID="322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.452020 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-94jzf"] Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.475148 4888 scope.go:117] "RemoveContainer" containerID="4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.529518 4888 scope.go:117] "RemoveContainer" containerID="6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d" Nov 24 02:41:03 crc kubenswrapper[4888]: E1124 02:41:03.530061 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d\": container with ID starting with 6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d not found: ID does not exist" containerID="6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.530115 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d"} err="failed to get container status \"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d\": rpc error: code = NotFound desc = could not find container \"6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d\": container with ID starting with 6b4b232c579315882fc0491de0b82371cf6878a805e707226c966bfc5683da5d not found: ID does not exist" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.530150 4888 scope.go:117] "RemoveContainer" containerID="322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095" Nov 24 02:41:03 crc kubenswrapper[4888]: E1124 02:41:03.530445 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095\": container with ID starting with 322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095 not found: ID does not exist" containerID="322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.530475 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095"} err="failed to get container status \"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095\": rpc error: code = NotFound desc = could not find container \"322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095\": container with ID starting with 322f5a3695d0d14da10c66582e4e54dcd533719f1a6de140d41eb6dd555e0095 not found: ID does not exist" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.530493 4888 scope.go:117] "RemoveContainer" containerID="4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff" Nov 24 02:41:03 crc kubenswrapper[4888]: E1124 02:41:03.530735 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff\": container with ID starting with 4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff not found: ID does not exist" containerID="4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff" Nov 24 02:41:03 crc kubenswrapper[4888]: I1124 02:41:03.530766 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff"} err="failed to get container status \"4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff\": rpc error: code = NotFound desc = could not find container \"4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff\": container with ID starting with 4a7e5c7b6493ff15de436c35b83c730bad7c50be524245d38d2603af222b4eff not found: ID does not exist" Nov 24 02:41:04 crc kubenswrapper[4888]: I1124 02:41:04.267757 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" path="/var/lib/kubelet/pods/dd2a6cd6-8d55-4247-83dd-c5b69975071b/volumes" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.085118 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lv57g_2b99942b-b619-4a73-bae0-b3cd34990e42/cluster-logging-operator/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.244699 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-gmbz6_61cf37cb-90a9-4b6a-8252-30a564f91b0d/collector/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.253718 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_38b36998-57a3-4b56-b3dd-16f2800890d8/loki-compactor/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.415050 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-4kw2j_9dfccf91-b0dd-46c1-a8a0-784d83237bbe/loki-distributor/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.461423 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-7m9r2_7000141c-3ebe-45cf-bdc1-36be274b4d8c/gateway/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.486438 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-7m9r2_7000141c-3ebe-45cf-bdc1-36be274b4d8c/opa/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.603005 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-hjsdr_8c68d551-242e-47b9-8ed7-d47f6f256d3f/gateway/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.636506 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-76846dc87d-hjsdr_8c68d551-242e-47b9-8ed7-d47f6f256d3f/opa/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.767900 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_e8d0d952-6592-40e8-a831-5e875e1f9d4b/loki-index-gateway/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.881759 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_158da35a-1b60-4ef6-b46a-fd8d3e8c5c51/loki-ingester/0.log" Nov 24 02:41:12 crc kubenswrapper[4888]: I1124 02:41:12.976184 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-69xbm_03bd362f-47ff-4c5c-8ae8-b60c0c87138b/loki-querier/0.log" Nov 24 02:41:13 crc kubenswrapper[4888]: I1124 02:41:13.027839 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-ntcth_2386f5e1-9382-4d87-b641-4372f630d9d4/loki-query-frontend/0.log" Nov 24 02:41:23 crc kubenswrapper[4888]: I1124 02:41:23.591890 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:41:23 crc kubenswrapper[4888]: I1124 02:41:23.592501 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:41:26 crc kubenswrapper[4888]: I1124 02:41:26.853031 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-77rp4_e710f559-91d6-47ce-bf7d-ee5717c276a8/kube-rbac-proxy/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.055840 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-77rp4_e710f559-91d6-47ce-bf7d-ee5717c276a8/controller/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.105916 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.312219 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.320840 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.321018 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.349420 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.497805 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.527146 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.558012 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.562671 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.724882 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-frr-files/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.742497 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-reloader/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.760998 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/controller/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.783579 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/cp-metrics/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.948929 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/frr-metrics/0.log" Nov 24 02:41:27 crc kubenswrapper[4888]: I1124 02:41:27.981603 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/kube-rbac-proxy-frr/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.028416 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/kube-rbac-proxy/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.213317 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/reloader/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.291050 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-zvfv2_43445b39-e7aa-4d6c-bc1a-760e001605b3/frr-k8s-webhook-server/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.431015 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f89595f9b-rs4gm_d8651fd8-33fc-4d15-b478-78921c04fef1/manager/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.664708 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69f6b76c88-dh8sn_007267a1-8af0-47c2-84e1-8fa533e3f5c7/webhook-server/0.log" Nov 24 02:41:28 crc kubenswrapper[4888]: I1124 02:41:28.739470 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2lxkv_2b50d2f6-eedb-4817-ba29-4b6e908aae96/kube-rbac-proxy/0.log" Nov 24 02:41:29 crc kubenswrapper[4888]: I1124 02:41:29.449468 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2lxkv_2b50d2f6-eedb-4817-ba29-4b6e908aae96/speaker/0.log" Nov 24 02:41:30 crc kubenswrapper[4888]: I1124 02:41:30.220539 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bdrm6_5b721b4f-a010-471e-a766-159936fc9379/frr/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.550497 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.733721 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.736684 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.773718 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.938060 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/extract/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.943829 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/pull/0.log" Nov 24 02:41:42 crc kubenswrapper[4888]: I1124 02:41:42.980837 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8hcdkm_ff873a77-87b3-42db-8bc1-acdcca37ef42/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.092791 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.269328 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.283559 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.298033 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.458286 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.469747 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/extract/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.488475 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e97v7x_e749cc09-5587-4044-878a-c3adbd64909a/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.597713 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.776931 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.815263 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.815519 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.986611 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/pull/0.log" Nov 24 02:41:43 crc kubenswrapper[4888]: I1124 02:41:43.998124 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/extract/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.009564 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v7snm_04e91586-5c1b-490b-ac59-7f0f162623cd/util/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.142853 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.304102 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.305678 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.328653 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.458121 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/util/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.465225 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/pull/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.524150 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fv2tss_25b7bdca-37de-4056-8f0f-781a398ad19e/extract/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.647235 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.836414 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.846922 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:41:44 crc kubenswrapper[4888]: I1124 02:41:44.886021 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.035635 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-content/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.040337 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/extract-utilities/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.321038 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.515165 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.565259 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.599241 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.782042 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-utilities/0.log" Nov 24 02:41:45 crc kubenswrapper[4888]: I1124 02:41:45.807449 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/extract-content/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.042101 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.157677 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qlhkx_cab99235-2ee5-487d-be36-3d78fff69946/registry-server/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.252266 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.253129 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.285053 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.485975 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/util/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.495149 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/pull/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.573349 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62zbh5_8210c0ef-1d11-4f4f-b05d-b584490ecd29/extract/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.742447 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nmmd_ef22eb11-250d-44e2-a2bf-7906d7817b8b/marketplace-operator/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.792865 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.964067 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.973937 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p29pb_6edb2361-460b-4aa7-8d4c-0338d630b1ca/registry-server/0.log" Nov 24 02:41:46 crc kubenswrapper[4888]: I1124 02:41:46.976797 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.015800 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.148353 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-content/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.154912 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/extract-utilities/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.224446 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.382419 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.394284 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zc5l_042dca76-9304-4100-87ff-a67fa19b4674/registry-server/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.409271 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.428629 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.592845 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-utilities/0.log" Nov 24 02:41:47 crc kubenswrapper[4888]: I1124 02:41:47.596222 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/extract-content/0.log" Nov 24 02:41:48 crc kubenswrapper[4888]: I1124 02:41:48.562939 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v97rx_d48aee78-3eb2-439e-9ae4-6d56a796ad1c/registry-server/0.log" Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.591377 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.591960 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.592010 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.592873 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.592926 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae" gracePeriod=600 Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.935581 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae" exitCode=0 Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.935792 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae"} Nov 24 02:41:53 crc kubenswrapper[4888]: I1124 02:41:53.936059 4888 scope.go:117] "RemoveContainer" containerID="ec13446c39d99e97ae9c31b9c72b59a91239d9114fc1aabee1610acc4e3478ec" Nov 24 02:41:54 crc kubenswrapper[4888]: I1124 02:41:54.951485 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerStarted","Data":"8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5"} Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.129187 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-x756x_85dc9246-ac2d-44f7-bda0-bdd6faa6a391/prometheus-operator/0.log" Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.312949 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5ffd7d97f9-8krz2_a11b0ce1-5305-4b79-a302-4839c19791e0/prometheus-operator-admission-webhook/0.log" Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.316568 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5ffd7d97f9-vgfxd_9a7b0bb7-bf4c-47f3-a39d-b05a758f50f4/prometheus-operator-admission-webhook/0.log" Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.462385 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-vj7xs_2c9cc7d0-818c-484c-96fd-c22c437348bc/operator/0.log" Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.488939 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-clzxk_a05abc8e-614d-4cb7-8eda-35f0698a756c/observability-ui-dashboards/0.log" Nov 24 02:42:00 crc kubenswrapper[4888]: I1124 02:42:00.654228 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-7j7tc_920040ee-4308-431b-9f6d-eb16563acb3f/perses-operator/0.log" Nov 24 02:42:14 crc kubenswrapper[4888]: I1124 02:42:14.170954 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/kube-rbac-proxy/0.log" Nov 24 02:42:14 crc kubenswrapper[4888]: I1124 02:42:14.229675 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-648dc49c56-gnzhc_d9fc74ad-5691-4d31-a9b8-9bab98d49121/manager/0.log" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.981268 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:42:52 crc kubenswrapper[4888]: E1124 02:42:52.982552 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.982570 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" Nov 24 02:42:52 crc kubenswrapper[4888]: E1124 02:42:52.982592 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="extract-utilities" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.982600 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="extract-utilities" Nov 24 02:42:52 crc kubenswrapper[4888]: E1124 02:42:52.982641 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="extract-content" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.982649 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="extract-content" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.982886 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2a6cd6-8d55-4247-83dd-c5b69975071b" containerName="registry-server" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.984627 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:52 crc kubenswrapper[4888]: I1124 02:42:52.998190 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.041598 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q8tw\" (UniqueName: \"kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.041681 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.041941 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.143612 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.143701 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q8tw\" (UniqueName: \"kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.143755 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.144233 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.144233 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.174454 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q8tw\" (UniqueName: \"kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw\") pod \"community-operators-rd5bc\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:53 crc kubenswrapper[4888]: I1124 02:42:53.312630 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:42:54 crc kubenswrapper[4888]: I1124 02:42:54.288851 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:42:54 crc kubenswrapper[4888]: I1124 02:42:54.674577 4888 generic.go:334] "Generic (PLEG): container finished" podID="410f7774-beef-486a-a3fb-6aaac148cc69" containerID="b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5" exitCode=0 Nov 24 02:42:54 crc kubenswrapper[4888]: I1124 02:42:54.674880 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerDied","Data":"b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5"} Nov 24 02:42:54 crc kubenswrapper[4888]: I1124 02:42:54.674912 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerStarted","Data":"18fe0496366e795e4b4460ed315b67d60ffcc3fef1de7f80ff180d734b225766"} Nov 24 02:42:54 crc kubenswrapper[4888]: I1124 02:42:54.677120 4888 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 02:42:55 crc kubenswrapper[4888]: I1124 02:42:55.687125 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerStarted","Data":"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa"} Nov 24 02:42:57 crc kubenswrapper[4888]: I1124 02:42:57.735873 4888 generic.go:334] "Generic (PLEG): container finished" podID="410f7774-beef-486a-a3fb-6aaac148cc69" containerID="1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa" exitCode=0 Nov 24 02:42:57 crc kubenswrapper[4888]: I1124 02:42:57.735988 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerDied","Data":"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa"} Nov 24 02:42:58 crc kubenswrapper[4888]: I1124 02:42:58.753283 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerStarted","Data":"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a"} Nov 24 02:42:58 crc kubenswrapper[4888]: I1124 02:42:58.778308 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rd5bc" podStartSLOduration=3.223624463 podStartE2EDuration="6.778285533s" podCreationTimestamp="2025-11-24 02:42:52 +0000 UTC" firstStartedPulling="2025-11-24 02:42:54.676871761 +0000 UTC m=+8277.259555805" lastFinishedPulling="2025-11-24 02:42:58.231532821 +0000 UTC m=+8280.814216875" observedRunningTime="2025-11-24 02:42:58.775310329 +0000 UTC m=+8281.357994373" watchObservedRunningTime="2025-11-24 02:42:58.778285533 +0000 UTC m=+8281.360969577" Nov 24 02:43:03 crc kubenswrapper[4888]: I1124 02:43:03.313566 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:03 crc kubenswrapper[4888]: I1124 02:43:03.314330 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:03 crc kubenswrapper[4888]: I1124 02:43:03.372582 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:03 crc kubenswrapper[4888]: I1124 02:43:03.902103 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:03 crc kubenswrapper[4888]: I1124 02:43:03.967856 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:43:05 crc kubenswrapper[4888]: I1124 02:43:05.858357 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rd5bc" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="registry-server" containerID="cri-o://405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a" gracePeriod=2 Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.456246 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.608197 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content\") pod \"410f7774-beef-486a-a3fb-6aaac148cc69\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.608403 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities\") pod \"410f7774-beef-486a-a3fb-6aaac148cc69\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.608515 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q8tw\" (UniqueName: \"kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw\") pod \"410f7774-beef-486a-a3fb-6aaac148cc69\" (UID: \"410f7774-beef-486a-a3fb-6aaac148cc69\") " Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.609230 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities" (OuterVolumeSpecName: "utilities") pod "410f7774-beef-486a-a3fb-6aaac148cc69" (UID: "410f7774-beef-486a-a3fb-6aaac148cc69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.615475 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw" (OuterVolumeSpecName: "kube-api-access-6q8tw") pod "410f7774-beef-486a-a3fb-6aaac148cc69" (UID: "410f7774-beef-486a-a3fb-6aaac148cc69"). InnerVolumeSpecName "kube-api-access-6q8tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.674134 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "410f7774-beef-486a-a3fb-6aaac148cc69" (UID: "410f7774-beef-486a-a3fb-6aaac148cc69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.711330 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.711376 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q8tw\" (UniqueName: \"kubernetes.io/projected/410f7774-beef-486a-a3fb-6aaac148cc69-kube-api-access-6q8tw\") on node \"crc\" DevicePath \"\"" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.711399 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410f7774-beef-486a-a3fb-6aaac148cc69-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.872261 4888 generic.go:334] "Generic (PLEG): container finished" podID="410f7774-beef-486a-a3fb-6aaac148cc69" containerID="405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a" exitCode=0 Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.872313 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerDied","Data":"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a"} Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.872290 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd5bc" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.872398 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd5bc" event={"ID":"410f7774-beef-486a-a3fb-6aaac148cc69","Type":"ContainerDied","Data":"18fe0496366e795e4b4460ed315b67d60ffcc3fef1de7f80ff180d734b225766"} Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.872428 4888 scope.go:117] "RemoveContainer" containerID="405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.892737 4888 scope.go:117] "RemoveContainer" containerID="1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa" Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.928473 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.941144 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rd5bc"] Nov 24 02:43:06 crc kubenswrapper[4888]: I1124 02:43:06.943417 4888 scope.go:117] "RemoveContainer" containerID="b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.003709 4888 scope.go:117] "RemoveContainer" containerID="405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a" Nov 24 02:43:07 crc kubenswrapper[4888]: E1124 02:43:07.004131 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a\": container with ID starting with 405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a not found: ID does not exist" containerID="405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.004161 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a"} err="failed to get container status \"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a\": rpc error: code = NotFound desc = could not find container \"405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a\": container with ID starting with 405848f84cdd0c4d728802f9f82bf48bb36429a613cb07f39936f27578fac89a not found: ID does not exist" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.004182 4888 scope.go:117] "RemoveContainer" containerID="1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa" Nov 24 02:43:07 crc kubenswrapper[4888]: E1124 02:43:07.004560 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa\": container with ID starting with 1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa not found: ID does not exist" containerID="1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.004579 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa"} err="failed to get container status \"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa\": rpc error: code = NotFound desc = could not find container \"1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa\": container with ID starting with 1c3d3f0035d7adcf24d3426eb0726843fd6a8cc54b3ae92f2167493c80711ffa not found: ID does not exist" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.004592 4888 scope.go:117] "RemoveContainer" containerID="b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5" Nov 24 02:43:07 crc kubenswrapper[4888]: E1124 02:43:07.004908 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5\": container with ID starting with b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5 not found: ID does not exist" containerID="b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5" Nov 24 02:43:07 crc kubenswrapper[4888]: I1124 02:43:07.004938 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5"} err="failed to get container status \"b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5\": rpc error: code = NotFound desc = could not find container \"b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5\": container with ID starting with b7470fd7bb1ae1b8a2a51b60be2505f9e4b1e8548efa57f2b60b2ba72aebbcb5 not found: ID does not exist" Nov 24 02:43:08 crc kubenswrapper[4888]: I1124 02:43:08.259985 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" path="/var/lib/kubelet/pods/410f7774-beef-486a-a3fb-6aaac148cc69/volumes" Nov 24 02:43:23 crc kubenswrapper[4888]: I1124 02:43:23.573160 4888 scope.go:117] "RemoveContainer" containerID="ae391068cb5f2fa738454c43623e3c843baddf40a2ff773ee6d9b88c9fead4cb" Nov 24 02:43:53 crc kubenswrapper[4888]: I1124 02:43:53.591068 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:43:53 crc kubenswrapper[4888]: I1124 02:43:53.591654 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:44:12 crc kubenswrapper[4888]: I1124 02:44:12.504566 4888 generic.go:334] "Generic (PLEG): container finished" podID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerID="8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3" exitCode=0 Nov 24 02:44:12 crc kubenswrapper[4888]: I1124 02:44:12.504783 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" event={"ID":"294d2afc-c188-4fc5-8f79-f14a0aecefa9","Type":"ContainerDied","Data":"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3"} Nov 24 02:44:12 crc kubenswrapper[4888]: I1124 02:44:12.506659 4888 scope.go:117] "RemoveContainer" containerID="8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3" Nov 24 02:44:13 crc kubenswrapper[4888]: I1124 02:44:13.227527 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x9v9r_must-gather-xlrb5_294d2afc-c188-4fc5-8f79-f14a0aecefa9/gather/0.log" Nov 24 02:44:23 crc kubenswrapper[4888]: I1124 02:44:23.591142 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:44:23 crc kubenswrapper[4888]: I1124 02:44:23.592988 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:44:23 crc kubenswrapper[4888]: I1124 02:44:23.691296 4888 scope.go:117] "RemoveContainer" containerID="64f99e22568a7d0ac64d91e232a5d04366542121657bc9ec6eccff010eb91f12" Nov 24 02:44:24 crc kubenswrapper[4888]: I1124 02:44:24.947434 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x9v9r/must-gather-xlrb5"] Nov 24 02:44:24 crc kubenswrapper[4888]: I1124 02:44:24.948056 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="copy" containerID="cri-o://eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd" gracePeriod=2 Nov 24 02:44:24 crc kubenswrapper[4888]: I1124 02:44:24.961889 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x9v9r/must-gather-xlrb5"] Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.410206 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x9v9r_must-gather-xlrb5_294d2afc-c188-4fc5-8f79-f14a0aecefa9/copy/0.log" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.410724 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.490981 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzgjh\" (UniqueName: \"kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh\") pod \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.491098 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output\") pod \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\" (UID: \"294d2afc-c188-4fc5-8f79-f14a0aecefa9\") " Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.497280 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh" (OuterVolumeSpecName: "kube-api-access-gzgjh") pod "294d2afc-c188-4fc5-8f79-f14a0aecefa9" (UID: "294d2afc-c188-4fc5-8f79-f14a0aecefa9"). InnerVolumeSpecName "kube-api-access-gzgjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.594535 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzgjh\" (UniqueName: \"kubernetes.io/projected/294d2afc-c188-4fc5-8f79-f14a0aecefa9-kube-api-access-gzgjh\") on node \"crc\" DevicePath \"\"" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.671658 4888 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x9v9r_must-gather-xlrb5_294d2afc-c188-4fc5-8f79-f14a0aecefa9/copy/0.log" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.672096 4888 generic.go:334] "Generic (PLEG): container finished" podID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerID="eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd" exitCode=143 Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.672136 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x9v9r/must-gather-xlrb5" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.672158 4888 scope.go:117] "RemoveContainer" containerID="eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.676689 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "294d2afc-c188-4fc5-8f79-f14a0aecefa9" (UID: "294d2afc-c188-4fc5-8f79-f14a0aecefa9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.695280 4888 scope.go:117] "RemoveContainer" containerID="8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.697623 4888 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/294d2afc-c188-4fc5-8f79-f14a0aecefa9-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.734016 4888 scope.go:117] "RemoveContainer" containerID="eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd" Nov 24 02:44:25 crc kubenswrapper[4888]: E1124 02:44:25.734478 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd\": container with ID starting with eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd not found: ID does not exist" containerID="eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.734516 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd"} err="failed to get container status \"eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd\": rpc error: code = NotFound desc = could not find container \"eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd\": container with ID starting with eb954b7f0399e011b7948bc7d335709a28ff01598de1f4ee83767cc0c8a12ffd not found: ID does not exist" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.734538 4888 scope.go:117] "RemoveContainer" containerID="8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3" Nov 24 02:44:25 crc kubenswrapper[4888]: E1124 02:44:25.734802 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3\": container with ID starting with 8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3 not found: ID does not exist" containerID="8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3" Nov 24 02:44:25 crc kubenswrapper[4888]: I1124 02:44:25.734859 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3"} err="failed to get container status \"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3\": rpc error: code = NotFound desc = could not find container \"8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3\": container with ID starting with 8a4a418b6532304e348f4be9d0a0d84aff9eb6527d68c9be2c357704f771c8c3 not found: ID does not exist" Nov 24 02:44:26 crc kubenswrapper[4888]: I1124 02:44:26.259081 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" path="/var/lib/kubelet/pods/294d2afc-c188-4fc5-8f79-f14a0aecefa9/volumes" Nov 24 02:44:53 crc kubenswrapper[4888]: I1124 02:44:53.592027 4888 patch_prober.go:28] interesting pod/machine-config-daemon-wvkcz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 02:44:53 crc kubenswrapper[4888]: I1124 02:44:53.592415 4888 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 02:44:53 crc kubenswrapper[4888]: I1124 02:44:53.592457 4888 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" Nov 24 02:44:53 crc kubenswrapper[4888]: I1124 02:44:53.593463 4888 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5"} pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 02:44:53 crc kubenswrapper[4888]: I1124 02:44:53.593583 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerName="machine-config-daemon" containerID="cri-o://8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" gracePeriod=600 Nov 24 02:44:53 crc kubenswrapper[4888]: E1124 02:44:53.720309 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:44:54 crc kubenswrapper[4888]: I1124 02:44:54.020472 4888 generic.go:334] "Generic (PLEG): container finished" podID="d17c94b3-864e-4594-8ed7-06819e60fa5e" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" exitCode=0 Nov 24 02:44:54 crc kubenswrapper[4888]: I1124 02:44:54.020526 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" event={"ID":"d17c94b3-864e-4594-8ed7-06819e60fa5e","Type":"ContainerDied","Data":"8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5"} Nov 24 02:44:54 crc kubenswrapper[4888]: I1124 02:44:54.020758 4888 scope.go:117] "RemoveContainer" containerID="2232accfc0923e1af5a445c4dbce091a668431df4dddf94dc1649028d7bedfae" Nov 24 02:44:54 crc kubenswrapper[4888]: I1124 02:44:54.022098 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:44:54 crc kubenswrapper[4888]: E1124 02:44:54.022749 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.151840 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf"] Nov 24 02:45:00 crc kubenswrapper[4888]: E1124 02:45:00.152875 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="gather" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.152891 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="gather" Nov 24 02:45:00 crc kubenswrapper[4888]: E1124 02:45:00.152915 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="extract-utilities" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.152924 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="extract-utilities" Nov 24 02:45:00 crc kubenswrapper[4888]: E1124 02:45:00.152954 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="registry-server" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.152963 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="registry-server" Nov 24 02:45:00 crc kubenswrapper[4888]: E1124 02:45:00.152982 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="copy" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.152990 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="copy" Nov 24 02:45:00 crc kubenswrapper[4888]: E1124 02:45:00.153007 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="extract-content" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.153017 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="extract-content" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.153266 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="gather" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.153285 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="410f7774-beef-486a-a3fb-6aaac148cc69" containerName="registry-server" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.153311 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="294d2afc-c188-4fc5-8f79-f14a0aecefa9" containerName="copy" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.154315 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.159732 4888 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.159732 4888 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.175939 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf"] Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.277407 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpbdx\" (UniqueName: \"kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.277804 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.277892 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.380097 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpbdx\" (UniqueName: \"kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.380209 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.380273 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.381185 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.387675 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.399347 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpbdx\" (UniqueName: \"kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx\") pod \"collect-profiles-29399205-wdttf\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.485570 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:00 crc kubenswrapper[4888]: I1124 02:45:00.953261 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf"] Nov 24 02:45:00 crc kubenswrapper[4888]: W1124 02:45:00.960140 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode19e02e7_c5cf_44ce_b65f_044d9a095e2c.slice/crio-19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4 WatchSource:0}: Error finding container 19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4: Status 404 returned error can't find the container with id 19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4 Nov 24 02:45:01 crc kubenswrapper[4888]: I1124 02:45:01.150212 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" event={"ID":"e19e02e7-c5cf-44ce-b65f-044d9a095e2c","Type":"ContainerStarted","Data":"5ef6d083ce710c1ce70fdb0416e56a8802464d53c05d37fc7eff13c6ad02a45f"} Nov 24 02:45:01 crc kubenswrapper[4888]: I1124 02:45:01.150259 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" event={"ID":"e19e02e7-c5cf-44ce-b65f-044d9a095e2c","Type":"ContainerStarted","Data":"19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4"} Nov 24 02:45:01 crc kubenswrapper[4888]: I1124 02:45:01.169969 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" podStartSLOduration=1.169953548 podStartE2EDuration="1.169953548s" podCreationTimestamp="2025-11-24 02:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 02:45:01.162186679 +0000 UTC m=+8403.744870723" watchObservedRunningTime="2025-11-24 02:45:01.169953548 +0000 UTC m=+8403.752637592" Nov 24 02:45:02 crc kubenswrapper[4888]: I1124 02:45:02.160713 4888 generic.go:334] "Generic (PLEG): container finished" podID="e19e02e7-c5cf-44ce-b65f-044d9a095e2c" containerID="5ef6d083ce710c1ce70fdb0416e56a8802464d53c05d37fc7eff13c6ad02a45f" exitCode=0 Nov 24 02:45:02 crc kubenswrapper[4888]: I1124 02:45:02.160840 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" event={"ID":"e19e02e7-c5cf-44ce-b65f-044d9a095e2c","Type":"ContainerDied","Data":"5ef6d083ce710c1ce70fdb0416e56a8802464d53c05d37fc7eff13c6ad02a45f"} Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.594612 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.763496 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpbdx\" (UniqueName: \"kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx\") pod \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.764134 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume\") pod \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.764175 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume\") pod \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\" (UID: \"e19e02e7-c5cf-44ce-b65f-044d9a095e2c\") " Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.765319 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume" (OuterVolumeSpecName: "config-volume") pod "e19e02e7-c5cf-44ce-b65f-044d9a095e2c" (UID: "e19e02e7-c5cf-44ce-b65f-044d9a095e2c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.770242 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx" (OuterVolumeSpecName: "kube-api-access-rpbdx") pod "e19e02e7-c5cf-44ce-b65f-044d9a095e2c" (UID: "e19e02e7-c5cf-44ce-b65f-044d9a095e2c"). InnerVolumeSpecName "kube-api-access-rpbdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.774449 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e19e02e7-c5cf-44ce-b65f-044d9a095e2c" (UID: "e19e02e7-c5cf-44ce-b65f-044d9a095e2c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.868211 4888 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.868275 4888 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:03 crc kubenswrapper[4888]: I1124 02:45:03.868294 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpbdx\" (UniqueName: \"kubernetes.io/projected/e19e02e7-c5cf-44ce-b65f-044d9a095e2c-kube-api-access-rpbdx\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.184392 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.184405 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399205-wdttf" event={"ID":"e19e02e7-c5cf-44ce-b65f-044d9a095e2c","Type":"ContainerDied","Data":"19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4"} Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.184937 4888 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19ab926fae33eccbfc49442a0e36f2adb33e7305f1087dc0c2f4802d806356b4" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.283730 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n"] Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.293673 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399160-rj55n"] Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.493053 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:04 crc kubenswrapper[4888]: E1124 02:45:04.493851 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19e02e7-c5cf-44ce-b65f-044d9a095e2c" containerName="collect-profiles" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.493880 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19e02e7-c5cf-44ce-b65f-044d9a095e2c" containerName="collect-profiles" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.494266 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19e02e7-c5cf-44ce-b65f-044d9a095e2c" containerName="collect-profiles" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.497209 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.518488 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.585306 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4hf\" (UniqueName: \"kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.585718 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.585802 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.687401 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4hf\" (UniqueName: \"kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.687582 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.687708 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.688015 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.688139 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.711993 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4hf\" (UniqueName: \"kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf\") pod \"certified-operators-gsln7\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:04 crc kubenswrapper[4888]: I1124 02:45:04.837129 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:05 crc kubenswrapper[4888]: I1124 02:45:05.418278 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:05 crc kubenswrapper[4888]: W1124 02:45:05.425291 4888 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14e46875_1b04_4a09_a78b_da351dbe1342.slice/crio-26b2cdc52502d9f2a4fb39d6c5d27e26b815f9023e98810f18a6dfb1c6871bfc WatchSource:0}: Error finding container 26b2cdc52502d9f2a4fb39d6c5d27e26b815f9023e98810f18a6dfb1c6871bfc: Status 404 returned error can't find the container with id 26b2cdc52502d9f2a4fb39d6c5d27e26b815f9023e98810f18a6dfb1c6871bfc Nov 24 02:45:06 crc kubenswrapper[4888]: I1124 02:45:06.207242 4888 generic.go:334] "Generic (PLEG): container finished" podID="14e46875-1b04-4a09-a78b-da351dbe1342" containerID="0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd" exitCode=0 Nov 24 02:45:06 crc kubenswrapper[4888]: I1124 02:45:06.207360 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerDied","Data":"0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd"} Nov 24 02:45:06 crc kubenswrapper[4888]: I1124 02:45:06.207540 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerStarted","Data":"26b2cdc52502d9f2a4fb39d6c5d27e26b815f9023e98810f18a6dfb1c6871bfc"} Nov 24 02:45:06 crc kubenswrapper[4888]: I1124 02:45:06.259653 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="665c431f-1a41-40c7-a486-c18c47be6d72" path="/var/lib/kubelet/pods/665c431f-1a41-40c7-a486-c18c47be6d72/volumes" Nov 24 02:45:07 crc kubenswrapper[4888]: I1124 02:45:07.220106 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerStarted","Data":"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893"} Nov 24 02:45:08 crc kubenswrapper[4888]: I1124 02:45:08.233100 4888 generic.go:334] "Generic (PLEG): container finished" podID="14e46875-1b04-4a09-a78b-da351dbe1342" containerID="5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893" exitCode=0 Nov 24 02:45:08 crc kubenswrapper[4888]: I1124 02:45:08.233152 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerDied","Data":"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893"} Nov 24 02:45:08 crc kubenswrapper[4888]: I1124 02:45:08.246340 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:45:08 crc kubenswrapper[4888]: E1124 02:45:08.246800 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:45:09 crc kubenswrapper[4888]: I1124 02:45:09.247766 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerStarted","Data":"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6"} Nov 24 02:45:09 crc kubenswrapper[4888]: I1124 02:45:09.267973 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gsln7" podStartSLOduration=2.747000765 podStartE2EDuration="5.267950229s" podCreationTimestamp="2025-11-24 02:45:04 +0000 UTC" firstStartedPulling="2025-11-24 02:45:06.210076155 +0000 UTC m=+8408.792760199" lastFinishedPulling="2025-11-24 02:45:08.731025599 +0000 UTC m=+8411.313709663" observedRunningTime="2025-11-24 02:45:09.263570576 +0000 UTC m=+8411.846254620" watchObservedRunningTime="2025-11-24 02:45:09.267950229 +0000 UTC m=+8411.850634273" Nov 24 02:45:14 crc kubenswrapper[4888]: I1124 02:45:14.837250 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:14 crc kubenswrapper[4888]: I1124 02:45:14.837299 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:14 crc kubenswrapper[4888]: I1124 02:45:14.884422 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:15 crc kubenswrapper[4888]: I1124 02:45:15.386336 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:15 crc kubenswrapper[4888]: I1124 02:45:15.441643 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:17 crc kubenswrapper[4888]: I1124 02:45:17.340140 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gsln7" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="registry-server" containerID="cri-o://11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6" gracePeriod=2 Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.181267 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.352847 4888 generic.go:334] "Generic (PLEG): container finished" podID="14e46875-1b04-4a09-a78b-da351dbe1342" containerID="11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6" exitCode=0 Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.352903 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerDied","Data":"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6"} Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.352934 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsln7" event={"ID":"14e46875-1b04-4a09-a78b-da351dbe1342","Type":"ContainerDied","Data":"26b2cdc52502d9f2a4fb39d6c5d27e26b815f9023e98810f18a6dfb1c6871bfc"} Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.352953 4888 scope.go:117] "RemoveContainer" containerID="11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.353048 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsln7" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.362884 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities\") pod \"14e46875-1b04-4a09-a78b-da351dbe1342\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.363049 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content\") pod \"14e46875-1b04-4a09-a78b-da351dbe1342\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.363228 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb4hf\" (UniqueName: \"kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf\") pod \"14e46875-1b04-4a09-a78b-da351dbe1342\" (UID: \"14e46875-1b04-4a09-a78b-da351dbe1342\") " Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.363866 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities" (OuterVolumeSpecName: "utilities") pod "14e46875-1b04-4a09-a78b-da351dbe1342" (UID: "14e46875-1b04-4a09-a78b-da351dbe1342"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.364385 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.370805 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf" (OuterVolumeSpecName: "kube-api-access-gb4hf") pod "14e46875-1b04-4a09-a78b-da351dbe1342" (UID: "14e46875-1b04-4a09-a78b-da351dbe1342"). InnerVolumeSpecName "kube-api-access-gb4hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.381625 4888 scope.go:117] "RemoveContainer" containerID="5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.426690 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14e46875-1b04-4a09-a78b-da351dbe1342" (UID: "14e46875-1b04-4a09-a78b-da351dbe1342"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.440337 4888 scope.go:117] "RemoveContainer" containerID="0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.466677 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14e46875-1b04-4a09-a78b-da351dbe1342-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.466726 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb4hf\" (UniqueName: \"kubernetes.io/projected/14e46875-1b04-4a09-a78b-da351dbe1342-kube-api-access-gb4hf\") on node \"crc\" DevicePath \"\"" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.493601 4888 scope.go:117] "RemoveContainer" containerID="11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6" Nov 24 02:45:18 crc kubenswrapper[4888]: E1124 02:45:18.494331 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6\": container with ID starting with 11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6 not found: ID does not exist" containerID="11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.494379 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6"} err="failed to get container status \"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6\": rpc error: code = NotFound desc = could not find container \"11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6\": container with ID starting with 11acc094645620f11e96de58b2f2047ee2dbb88c1fbc479a62cc1faa9bd73ce6 not found: ID does not exist" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.494408 4888 scope.go:117] "RemoveContainer" containerID="5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893" Nov 24 02:45:18 crc kubenswrapper[4888]: E1124 02:45:18.495958 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893\": container with ID starting with 5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893 not found: ID does not exist" containerID="5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.495993 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893"} err="failed to get container status \"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893\": rpc error: code = NotFound desc = could not find container \"5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893\": container with ID starting with 5712eb725a5deaf3af1e2c92a4bfb2da14f70dc5463e397d87c823e55499c893 not found: ID does not exist" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.496017 4888 scope.go:117] "RemoveContainer" containerID="0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd" Nov 24 02:45:18 crc kubenswrapper[4888]: E1124 02:45:18.496378 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd\": container with ID starting with 0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd not found: ID does not exist" containerID="0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.496414 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd"} err="failed to get container status \"0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd\": rpc error: code = NotFound desc = could not find container \"0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd\": container with ID starting with 0e40aa852e2227694ccf7803c122c664c38f6f8771d3b7b9994586fdca61b6dd not found: ID does not exist" Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.690435 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:18 crc kubenswrapper[4888]: I1124 02:45:18.701618 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gsln7"] Nov 24 02:45:20 crc kubenswrapper[4888]: I1124 02:45:20.245877 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:45:20 crc kubenswrapper[4888]: E1124 02:45:20.246682 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:45:20 crc kubenswrapper[4888]: I1124 02:45:20.259424 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" path="/var/lib/kubelet/pods/14e46875-1b04-4a09-a78b-da351dbe1342/volumes" Nov 24 02:45:23 crc kubenswrapper[4888]: I1124 02:45:23.753182 4888 scope.go:117] "RemoveContainer" containerID="b43b69fa6cad4c49cdb9666706072be630277ffeac37049dea6ee00bed2879cd" Nov 24 02:45:35 crc kubenswrapper[4888]: I1124 02:45:35.245707 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:45:35 crc kubenswrapper[4888]: E1124 02:45:35.247126 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:45:49 crc kubenswrapper[4888]: I1124 02:45:49.245052 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:45:49 crc kubenswrapper[4888]: E1124 02:45:49.245947 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:46:01 crc kubenswrapper[4888]: I1124 02:46:01.247157 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:46:01 crc kubenswrapper[4888]: E1124 02:46:01.248515 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:46:15 crc kubenswrapper[4888]: I1124 02:46:15.248624 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:46:15 crc kubenswrapper[4888]: E1124 02:46:15.249893 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:46:29 crc kubenswrapper[4888]: I1124 02:46:29.246411 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:46:29 crc kubenswrapper[4888]: E1124 02:46:29.247785 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:46:40 crc kubenswrapper[4888]: I1124 02:46:40.246301 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:46:40 crc kubenswrapper[4888]: E1124 02:46:40.247649 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:46:51 crc kubenswrapper[4888]: I1124 02:46:51.246927 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:46:51 crc kubenswrapper[4888]: E1124 02:46:51.248020 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:05 crc kubenswrapper[4888]: I1124 02:47:05.245779 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:47:05 crc kubenswrapper[4888]: E1124 02:47:05.246678 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:17 crc kubenswrapper[4888]: I1124 02:47:17.245998 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:47:17 crc kubenswrapper[4888]: E1124 02:47:17.246939 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:30 crc kubenswrapper[4888]: I1124 02:47:30.246063 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:47:30 crc kubenswrapper[4888]: E1124 02:47:30.246988 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.815766 4888 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:40 crc kubenswrapper[4888]: E1124 02:47:40.817921 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="extract-utilities" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.817960 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="extract-utilities" Nov 24 02:47:40 crc kubenswrapper[4888]: E1124 02:47:40.818001 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="registry-server" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.818018 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="registry-server" Nov 24 02:47:40 crc kubenswrapper[4888]: E1124 02:47:40.818140 4888 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="extract-content" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.818164 4888 state_mem.go:107] "Deleted CPUSet assignment" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="extract-content" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.818728 4888 memory_manager.go:354] "RemoveStaleState removing state" podUID="14e46875-1b04-4a09-a78b-da351dbe1342" containerName="registry-server" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.825885 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.840733 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.966862 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl72x\" (UniqueName: \"kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.967374 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:40 crc kubenswrapper[4888]: I1124 02:47:40.967447 4888 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.069230 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.069288 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.069405 4888 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl72x\" (UniqueName: \"kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.070190 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.070210 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.088912 4888 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl72x\" (UniqueName: \"kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x\") pod \"redhat-marketplace-t7bfr\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.149000 4888 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.245068 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:47:41 crc kubenswrapper[4888]: E1124 02:47:41.245340 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:41 crc kubenswrapper[4888]: I1124 02:47:41.673462 4888 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:42 crc kubenswrapper[4888]: I1124 02:47:42.241967 4888 generic.go:334] "Generic (PLEG): container finished" podID="7b7a936e-ce4b-4074-ae41-497f97176974" containerID="0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329" exitCode=0 Nov 24 02:47:42 crc kubenswrapper[4888]: I1124 02:47:42.242028 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerDied","Data":"0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329"} Nov 24 02:47:42 crc kubenswrapper[4888]: I1124 02:47:42.242342 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerStarted","Data":"42e7a5382f25fa52e59f517fc5b5f37b703cb377eabef393d449642b07240b00"} Nov 24 02:47:43 crc kubenswrapper[4888]: I1124 02:47:43.259241 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerStarted","Data":"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5"} Nov 24 02:47:44 crc kubenswrapper[4888]: I1124 02:47:44.277715 4888 generic.go:334] "Generic (PLEG): container finished" podID="7b7a936e-ce4b-4074-ae41-497f97176974" containerID="b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5" exitCode=0 Nov 24 02:47:44 crc kubenswrapper[4888]: I1124 02:47:44.278041 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerDied","Data":"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5"} Nov 24 02:47:45 crc kubenswrapper[4888]: I1124 02:47:45.292661 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerStarted","Data":"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a"} Nov 24 02:47:45 crc kubenswrapper[4888]: I1124 02:47:45.324672 4888 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7bfr" podStartSLOduration=2.859374264 podStartE2EDuration="5.324647273s" podCreationTimestamp="2025-11-24 02:47:40 +0000 UTC" firstStartedPulling="2025-11-24 02:47:42.245084756 +0000 UTC m=+8564.827768810" lastFinishedPulling="2025-11-24 02:47:44.710357745 +0000 UTC m=+8567.293041819" observedRunningTime="2025-11-24 02:47:45.318664504 +0000 UTC m=+8567.901348578" watchObservedRunningTime="2025-11-24 02:47:45.324647273 +0000 UTC m=+8567.907331357" Nov 24 02:47:51 crc kubenswrapper[4888]: I1124 02:47:51.149772 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:51 crc kubenswrapper[4888]: I1124 02:47:51.150416 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:51 crc kubenswrapper[4888]: I1124 02:47:51.245157 4888 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:51 crc kubenswrapper[4888]: I1124 02:47:51.412735 4888 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:51 crc kubenswrapper[4888]: I1124 02:47:51.499704 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:52 crc kubenswrapper[4888]: I1124 02:47:52.245635 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:47:52 crc kubenswrapper[4888]: E1124 02:47:52.246769 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:47:53 crc kubenswrapper[4888]: I1124 02:47:53.403978 4888 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7bfr" podUID="7b7a936e-ce4b-4074-ae41-497f97176974" containerName="registry-server" containerID="cri-o://46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a" gracePeriod=2 Nov 24 02:47:53 crc kubenswrapper[4888]: I1124 02:47:53.997140 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.112679 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content\") pod \"7b7a936e-ce4b-4074-ae41-497f97176974\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.112798 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl72x\" (UniqueName: \"kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x\") pod \"7b7a936e-ce4b-4074-ae41-497f97176974\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.113950 4888 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities\") pod \"7b7a936e-ce4b-4074-ae41-497f97176974\" (UID: \"7b7a936e-ce4b-4074-ae41-497f97176974\") " Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.115590 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities" (OuterVolumeSpecName: "utilities") pod "7b7a936e-ce4b-4074-ae41-497f97176974" (UID: "7b7a936e-ce4b-4074-ae41-497f97176974"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.121541 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x" (OuterVolumeSpecName: "kube-api-access-jl72x") pod "7b7a936e-ce4b-4074-ae41-497f97176974" (UID: "7b7a936e-ce4b-4074-ae41-497f97176974"). InnerVolumeSpecName "kube-api-access-jl72x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.153947 4888 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b7a936e-ce4b-4074-ae41-497f97176974" (UID: "7b7a936e-ce4b-4074-ae41-497f97176974"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.217382 4888 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.217436 4888 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl72x\" (UniqueName: \"kubernetes.io/projected/7b7a936e-ce4b-4074-ae41-497f97176974-kube-api-access-jl72x\") on node \"crc\" DevicePath \"\"" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.217456 4888 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b7a936e-ce4b-4074-ae41-497f97176974-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.415997 4888 generic.go:334] "Generic (PLEG): container finished" podID="7b7a936e-ce4b-4074-ae41-497f97176974" containerID="46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a" exitCode=0 Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.416055 4888 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7bfr" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.416076 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerDied","Data":"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a"} Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.417471 4888 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7bfr" event={"ID":"7b7a936e-ce4b-4074-ae41-497f97176974","Type":"ContainerDied","Data":"42e7a5382f25fa52e59f517fc5b5f37b703cb377eabef393d449642b07240b00"} Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.417512 4888 scope.go:117] "RemoveContainer" containerID="46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.453534 4888 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.458977 4888 scope.go:117] "RemoveContainer" containerID="b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.465203 4888 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7bfr"] Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.511731 4888 scope.go:117] "RemoveContainer" containerID="0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.545318 4888 scope.go:117] "RemoveContainer" containerID="46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a" Nov 24 02:47:54 crc kubenswrapper[4888]: E1124 02:47:54.546030 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a\": container with ID starting with 46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a not found: ID does not exist" containerID="46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.546068 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a"} err="failed to get container status \"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a\": rpc error: code = NotFound desc = could not find container \"46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a\": container with ID starting with 46818e8b9f7e73e1d1d0a2f23cfb6a9e722018567d2dfec5320e814ab730555a not found: ID does not exist" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.546098 4888 scope.go:117] "RemoveContainer" containerID="b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5" Nov 24 02:47:54 crc kubenswrapper[4888]: E1124 02:47:54.546425 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5\": container with ID starting with b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5 not found: ID does not exist" containerID="b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.546459 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5"} err="failed to get container status \"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5\": rpc error: code = NotFound desc = could not find container \"b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5\": container with ID starting with b7c8f42264018891b9c2c1a2dd12fb5309adfe251d5752e91dd24bdf2169b1d5 not found: ID does not exist" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.546477 4888 scope.go:117] "RemoveContainer" containerID="0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329" Nov 24 02:47:54 crc kubenswrapper[4888]: E1124 02:47:54.546726 4888 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329\": container with ID starting with 0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329 not found: ID does not exist" containerID="0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329" Nov 24 02:47:54 crc kubenswrapper[4888]: I1124 02:47:54.546753 4888 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329"} err="failed to get container status \"0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329\": rpc error: code = NotFound desc = could not find container \"0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329\": container with ID starting with 0f5fd882c4764eba43151f6c6014e28cb5d93c938ad320008577cf075c2d8329 not found: ID does not exist" Nov 24 02:47:56 crc kubenswrapper[4888]: I1124 02:47:56.268340 4888 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b7a936e-ce4b-4074-ae41-497f97176974" path="/var/lib/kubelet/pods/7b7a936e-ce4b-4074-ae41-497f97176974/volumes" Nov 24 02:48:04 crc kubenswrapper[4888]: I1124 02:48:04.245878 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:48:04 crc kubenswrapper[4888]: E1124 02:48:04.249207 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:48:19 crc kubenswrapper[4888]: I1124 02:48:19.245990 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:48:19 crc kubenswrapper[4888]: E1124 02:48:19.247073 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:48:30 crc kubenswrapper[4888]: I1124 02:48:30.249207 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:48:30 crc kubenswrapper[4888]: E1124 02:48:30.250423 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:48:44 crc kubenswrapper[4888]: I1124 02:48:44.246863 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:48:44 crc kubenswrapper[4888]: E1124 02:48:44.248189 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:48:55 crc kubenswrapper[4888]: I1124 02:48:55.245613 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:48:55 crc kubenswrapper[4888]: E1124 02:48:55.246684 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:49:06 crc kubenswrapper[4888]: I1124 02:49:06.246037 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:49:06 crc kubenswrapper[4888]: E1124 02:49:06.248420 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" Nov 24 02:49:18 crc kubenswrapper[4888]: I1124 02:49:18.268197 4888 scope.go:117] "RemoveContainer" containerID="8065440f20df16e8d47689c76922f752a6240f70ec0b527438bd941db6310fa5" Nov 24 02:49:18 crc kubenswrapper[4888]: E1124 02:49:18.269626 4888 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wvkcz_openshift-machine-config-operator(d17c94b3-864e-4594-8ed7-06819e60fa5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-wvkcz" podUID="d17c94b3-864e-4594-8ed7-06819e60fa5e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110743670024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110743671017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110722410016476 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110722410015446 5ustar corecore